Wav2vec2 Large 100k Voxpopuli
A speech recognition model pre-trained on 100,000 hours of unlabeled data from the VoxPopuli corpus, supporting multilingual speech representation learning
Downloads 2,218
Release Time : 3/2/2022
Model Overview
A large pre-trained Wav2Vec2 model developed by Facebook, focusing on learning speech structures from raw audio, suitable for speech recognition tasks
Model Features
Multilingual support
Trained on the VoxPopuli multilingual corpus, supporting speech processing in multiple languages
Unsupervised pre-training
Pre-trained with 100,000 hours of unlabeled speech data to learn general speech representations
Fine-tunable architecture
The model is designed for easy fine-tuning for specific languages or tasks
Model Capabilities
Speech representation learning
Automatic speech recognition
Multilingual speech processing
Use Cases
Speech technology
Speech recognition system
Build language-specific speech recognition systems by fine-tuning the model
Speech data analysis
Used for analyzing and processing speech content
Educational technology
Language learning applications
Build tools for speech assessment and pronunciation correction
Featured Recommended AI Models
Š 2025AIbase