H

Hubert Base Ls960

Developed by facebook
HuBERT is a self-supervised speech representation learning model that learns speech features through BERT-like prediction loss, suitable for tasks such as speech recognition.
Downloads 406.60k
Release Time : 3/2/2022

Model Overview

HuBERT (Hidden Unit BERT) is a self-supervised speech representation learning method that provides target labels for BERT-like prediction loss through an offline clustering step. The model is pre-trained on speech audio sampled at 16kHz and is suitable for tasks such as speech recognition, generation, and compression.

Model Features

Self-supervised learning
Provides target labels through unsupervised clustering steps, enabling speech representation learning without the need for large amounts of labeled data.
Efficient speech representation
Combines acoustic and language models on continuous input to learn efficient speech feature representations.
High performance
Outperforms or matches the state-of-the-art wav2vec 2.0 model on Librispeech and Libri-light benchmarks.

Model Capabilities

Speech representation learning
Speech recognition
Speech generation
Speech compression

Use Cases

Speech recognition
Automatic speech transcription
Converts speech audio into text, suitable for scenarios such as meeting minutes and subtitle generation.
Performs excellently on the Librispeech test set, reducing relative word error rate by 13-19%.
Speech generation
Speech synthesis
Combines with other models to generate natural speech.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase