M

Mert Base

Developed by yangwang825
MERT is an acoustic music understanding model based on self-supervised learning, using pseudo-labels provided by a teacher model for pre-training.
Downloads 26
Release Time : 8/6/2023

Model Overview

The MERT model focuses on audio classification tasks, particularly in the field of music understanding. It enhances model performance by introducing a teacher model to provide pseudo-labels in masked language modeling (MLM)-style acoustic pre-training.

Model Features

Self-supervised Pre-training
Employs large-scale self-supervised training methods to learn effective features without requiring extensive labeled data.
Teacher Model Guidance
Introduces a teacher model during pre-training to provide pseudo-labels, improving training effectiveness.
Multi-sample Rate Support
Capable of processing audio inputs with different sample rates (16kHz-44.1kHz).

Model Capabilities

Audio Feature Extraction
Music Classification
Acoustic Signal Processing

Use Cases

Music Analysis
Music Genre Classification
Automatically classify music clips by genre.
Music Emotion Recognition
Identify the emotional type expressed in music.
Audio Processing
Audio Feature Extraction
Extract high-level feature representations from audio.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase