Barthez
BARThez is a French sequence-to-sequence pre-trained model based on the BART architecture, particularly suitable for generative tasks such as abstractive summarization.
Downloads 1,487
Release Time : 3/2/2022
Model Overview
BARThez is a specialized French sequence-to-sequence pre-trained model that reconstructs corrupted input sentences during pre-training, utilizing a 66GB raw French text corpus. Unlike existing BERT-based French models, BARThez is especially suited for generative tasks as it pre-trains both the encoder and the decoder.
Model Features
Optimized for Generative Tasks
BARThez pre-trains both the encoder and decoder, making it particularly suitable for generative tasks like abstractive summarization.
Large-Scale Pre-training
Pre-trained on a 66GB raw French text corpus, covering a wide range of linguistic phenomena.
Multi-Version Support
Offers both base (BARThez) and large (mBARThez) versions to meet different needs.
Model Capabilities
Text Generation
Abstractive Summarization
Masked Language Modeling
Use Cases
Natural Language Processing
French Text Summarization
Use BARThez to generate abstractive summaries of French texts.
Excellent performance in generative tasks.
Masked Language Modeling
Use BARThez to fill in masked parts of sentences.
Accurately predicts missing vocabulary.
Featured Recommended AI Models
Š 2025AIbase