Model Overview
Model Features
Model Capabilities
Use Cases
đ LeBenchmark: wav2vec2 large model trained on 3K hours of French speech
LeBenchmark offers an ensemble of pretrained wav2vec2 models on various French datasets, which include spontaneous, read, and broadcasted speech. It has two versions. The later version (LeBenchmark 2.0) is an extended one compared to the first version, both in terms of the number of pre - trained SSL models and the number of downstream tasks. For more information on the different benchmarks for evaluating the wav2vec2 models, please refer to our paper at: LeBenchmark 2.0: a Standardized, Replicable and Enhanced Framework for Self - supervised Representations of French Speech
⨠Features
- Provides multiple wav2vec2 models trained on different French speech corpora.
- The models are available under the HuggingFace organization.
- The models can be reused under the Apache - 2.0 license.
- Can be fine - tuned for ASR with CTC using Fairseq tools.
- Can be integrated into the SpeechBrain toolkit for various speech - related tasks.
đĻ Installation
No installation steps are provided in the original document, so this section is skipped.
đģ Usage Examples
No code examples are provided in the original document, so this section is skipped.
đ Documentation
Model and data descriptions
We release four different models that can be found under our HuggingFace organization. Four different wav2vec2 architectures Light, Base, Large and xLarge are coupled with our small (1K), medium (3K), large (7K), and extra large (14K) corpus. In short:
Lebenchmark 2.0:
- [wav2vec2 - FR - 14K - xlarge](https://huggingface.co/LeBenchmark/wav2vec2 - FR - 14K - xlarge): xLarge wav2vec2 trained on 14K hours of French speech (5.4K Males / 2.4K Females / 6.8K unknown).
- [wav2vec2 - FR - 14K - large](https://huggingface.co/LeBenchmark/wav2vec2 - FR - 14K - large): Large wav2vec2 trained on 14K hours of French speech (5.4K Males / 2.4K Females / 6.8K unknown).
- [wav2vec2 - FR - 14K - light](https://huggingface.co/LeBenchmark/wav2vec2 - FR - 14K - light): Light wav2vec2 trained on 14K hours of French speech (5.4K Males / 2.4K Females / 6.8K unknown).
Lebenchmark:
- [wav2vec2 - FR - 7K - large](https://huggingface.co/LeBenchmark/wav2vec2 - FR - 7K - large): Large wav2vec2 trained on 7.6K hours of French speech (1.8K Males / 1.0K Females / 4.8K unknown).
- [wav2vec2 - FR - 7K - base](https://huggingface.co/LeBenchmark/wav2vec2 - FR - 7K - base): Base wav2vec2 trained on 7.6K hours of French speech (1.8K Males / 1.0K Females / 4.8K unknown).
- [wav2vec2 - FR - 3K - large](https://huggingface.co/LeBenchmark/wav2vec2 - FR - 3K - large): Large wav2vec2 trained on 2.9K hours of French speech (1.8K Males / 1.0K Females / 0.1K unknown).
- [wav2vec2 - FR - 3K - base](https://huggingface.co/LeBenchmark/wav2vec2 - FR - 3K - base): Base wav2vec2 trained on 2.9K hours of French speech (1.8K Males / 1.0K Females / 0.1K unknown).
- [wav2vec2 - FR - 2.6K - base](https://huggingface.co/LeBenchmark/wav2vec2 - FR - 2.6K - base): Base wav2vec2 trained on 2.6K hours of French speech (no spontaneous speech).
- [wav2vec2 - FR - 1K - large](https://huggingface.co/LeBenchmark/wav2vec2 - FR - 1K - large): Large wav2vec2 trained on 1K hours of French speech (0.5K Males / 0.5K Females).
- [wav2vec2 - FR - 1K - base](https://huggingface.co/LeBenchmark/wav2vec2 - FR - 1K - base): Base wav2vec2 trained on 1K hours of French speech (0.5K Males / 0.5K Females).
Intended uses & limitations
Pretrained wav2vec2 models are distributed under the Apache - 2.0 license. Hence, they can be reused extensively without strict limitations. However, benchmarks and data may be linked to corpora that are not completely open - sourced.
Fine - tune with Fairseq for ASR with CTC
As our wav2vec2 models were trained with Fairseq, they can be used in the different tools that Fairseq provides to fine - tune the model for ASR with CTC. The full procedure has been nicely summarized in [this blogpost](https://huggingface.co/blog/fine - tune - wav2vec2 - english).
â ī¸ Important Note
Due to the nature of CTC, speech - to - text results aren't expected to be state - of - the - art. Moreover, future features might appear depending on the involvement of Fairseq and HuggingFace on this part.
Integrate to SpeechBrain for ASR, Speaker, Source Separation ...
Pretrained wav2vec models recently gained in popularity. At the same time, SpeechBrain toolkit came out, proposing a new and simpler way of dealing with state - of - the - art speech & deep - learning technologies.
While it currently is in beta, SpeechBrain offers two different ways of nicely integrating wav2vec2 models that were trained with Fairseq i.e our LeBenchmark models!
- Extract wav2vec2 features on - the - fly (with a frozen wav2vec2 encoder) to be combined with any speech - related architecture. Examples are: E2E ASR with CTC+Att+Language Models; Speaker Recognition or Verification, Source Separation ...
- Experimental: To fully benefit from wav2vec2, the best solution remains to fine - tune the model while you train your downstream task. This is very simply allowed within SpeechBrain as just a flag needs to be turned on. Thus, our wav2vec2 models can be fine - tuned while training your favorite ASR pipeline or Speaker Recognizer.
đĄ Usage Tip
If interested, simply follow this tutorial
đ§ Technical Details
No specific technical details (more than 50 words) are provided in the original document, so this section is skipped.
đ License
Pretrained wav2vec2 models are distributed under the Apache - 2.0 license.
Referencing LeBenchmark
@misc{parcollet2023lebenchmark,
title={LeBenchmark 2.0: a Standardized, Replicable and Enhanced Framework for Self - supervised Representations of French Speech},
author={Titouan Parcollet and Ha Nguyen and Solene Evain and Marcely Zanon Boito and Adrien Pupier and Salima Mdhaffar and Hang Le and Sina Alisamir and Natalia Tomashenko and Marco Dinarelli and Shucong Zhang and Alexandre Allauzen and Maximin Coavoux and Yannick Esteve and Mickael Rouvier and Jerome Goulian and Benjamin Lecouteux and Francois Portet and Solange Rossato and Fabien Ringeval and Didier Schwab and Laurent Besacier},
year={2023},
eprint={2309.05472},
archivePrefix={arXiv},
primaryClass={cs.CL}
}

