Mgpt
mGPT is a multilingual generation model pretrained on the mC4 dataset, supporting 101 languages, using a Transformer architecture similar to GPT-2.
Downloads 147
Release Time : 3/2/2022
Model Overview
mGPT is a multilingual generation model based on the Transformer architecture, pretrained on the mC4 dataset of 101 languages through causal language modeling objectives. Suitable for text generation and prompt-based downstream task adaptation.
Model Features
Multilingual support
Pretraining covers 101 languages, offering broad language coverage
Unsupervised pretraining
Trained using raw text only, without the need for manually annotated data
Shared tokenizer
Uses the same tokenizer and vocabulary as mT5, facilitating multi-model collaboration
Model Capabilities
Multilingual text generation
Prompt-based task adaptation
Long text generation (up to 1024 tokens)
Use Cases
Text generation
Multilingual content creation
Generate coherent text content in different languages
Downstream task adaptation
Prompt-based task execution
Adapt to various NLP tasks by designing prompts
Featured Recommended AI Models
Š 2025AIbase