H

Hubert Xlarge Ll60k

Developed by facebook
Hubert is a self-supervised learning-based speech representation model that learns joint acoustic and linguistic representations of speech through BERT-like predictive loss.
Downloads 3,874
Release Time : 3/2/2022

Model Overview

This model is pre-trained on 16kHz sampled speech audio and is suitable for various downstream speech tasks such as automatic speech recognition and speaker recognition.

Model Features

Self-Supervised Learning
Learns speech representations through BERT-like predictive loss without requiring large amounts of labeled data.
Multi-Round Clustering Iteration
Uses two rounds of clustering iteration to optimize model performance and improve representation quality.
Joint Representation Learning
Simultaneously learns joint representations of acoustic and linguistic models, enhancing downstream task performance.

Model Capabilities

Speech representation extraction
Automatic speech recognition
Speaker recognition
Intent classification
Emotion recognition

Use Cases

Speech Processing
Automatic Speech Recognition
Converts speech into text
Achieves or surpasses the performance of wav2vec 2.0 on Librispeech and Libri-light benchmarks
Speaker Recognition
Identifies the speaker's identity in speech
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase