H

Hubert Large Ll60k

Developed by facebook
HuBERT is a self-supervised speech representation learning model that provides aligned target labels for BERT-like prediction loss through offline clustering steps, suitable for speech recognition, generation, and compression tasks.
Downloads 30.99k
Release Time : 3/2/2022

Model Overview

HuBERT (Hidden Unit BERT) is a self-supervised speech representation learning framework that excels in speech recognition tasks through innovative masked prediction loss and clustering methods.

Model Features

Self-supervised Learning
Pre-training without labeled data, learning speech representations through innovative masked prediction loss
Two-stage Clustering
Stably generates prediction target labels through initial k-means clustering and iterative optimization
Efficient Representation
Outperforms wav2vec 2.0 on Librispeech and Libri-light benchmarks
Large-scale Training
Supports training data scales ranging from 10 minutes to 60,000 hours

Model Capabilities

Speech representation learning
Speech recognition
Speech generation
Speech compression

Use Cases

Speech Technology
Speech Recognition System
Build high-accuracy speech recognition systems by fine-tuning the HuBERT model
Achieves 13-19% reduction in word error rate on the Librispeech test set
Speech Synthesis Frontend
Improve frontend processing of speech synthesis systems using learned speech representations
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase