A

Audio Magnet Medium

Developed by facebook
MAGNeT is a non-autoregressive Transformer-based text-to-music and sound effects generation model capable of producing high-quality audio samples from text descriptions.
Downloads 435
Release Time : 1/10/2024

Model Overview

MAGNeT employs a masked generative non-autoregressive Transformer architecture, trained on a 32kHz EnCodec tokenizer using four 50Hz sampled codebooks. The model generates all four codebooks through a single non-autoregressive Transformer without requiring semantic token conditioning or cascaded models.

Model Features

Single non-autoregressive Transformer architecture
Generates all codebooks through a single Transformer without cascaded models or semantic token conditioning.
High-quality audio generation
Produces high-quality music and sound effects samples from text descriptions.
Multi-codebook support
Trained on a 32kHz EnCodec tokenizer using four 50Hz sampled codebooks.

Model Capabilities

Text-to-music generation
Text-to-sound effects generation
High-quality audio sample generation

Use Cases

AI music generation research
Music composition assistance
Generate upbeat rock or energetic electronic dance music from text descriptions.
Produces high-quality music samples
Machine learning enthusiast exploration
Generative model capability exploration
Explore the application of non-autoregressive Transformers in audio generation.
Understand model performance in audio generation
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase