đ đ˛đŗ Mongol mGPT 1.3B
Mongol mGPT 1.3B is a language model specifically designed for the Mongol language. As the name suggests, this model has 1.3 billion parameters.
The Mongol language belongs to the Mongolic language family. It's a language with a long - standing history and is spoken by approximately 5.7 million people. Here are some key facts about it:
- It is the official language of Mongolia.
- In Mongolia, it uses the Cyrillic script, while the traditional Mongolian script is still in use in other regions.
- It has a rich history closely associated with the Mongol Empire and historical figures like Genghis Khan.
đ Quick Start
The provided README doesn't have specific quick - start steps. If you want to use this model, you can refer to the Hugging Face documentation related to mGPT models for general guidance on model loading and inference.
⨠Features
- Specifically tailored for the Mongol language, enabling more accurate language processing for Mongol - related tasks.
- Derived from a well - trained base model, which benefits from pre - training on a diverse set of languages.
đĻ Installation
The original README doesn't contain installation steps. You may follow the general installation process for Hugging Face models. Usually, you can use the transformers
library in Python:
pip install transformers
đ§ Technical Details
This model is one of the derivatives of the base [mGPT - XL (1.3B)](https://huggingface.co/ai - forever/mGPT) model. The base model was initially trained on 61 languages from 25 language families, using Wikipedia and the C4 corpus.
We discovered additional data for 23 languages, most of which are considered minor languages. Then, we decided to further fine - tune the base model. Mongol mGPT 1.3B was trained for an additional 50,000 steps with a batch size of 4 and a context window of 2048 tokens on 1 A100 GPU.
The final perplexity of this model on the validation set is 4.35.
Chart of the training loss and perplexity:

đ Documentation
Other mGPT - 1.3B models
- [đĻđ˛ mGPT - 1.3B Armenian](https://huggingface.co/ai - forever/mGPT - 1.3B - armenian)
- [đĻđŋ mGPT - 1.3B Azerbaijan](https://huggingface.co/ai - forever/mGPT - 1.3B - azerbaijan)
- [đ¯ mGPT - 1.3B Bashkir](https://huggingface.co/ai - forever/mGPT - 1.3B - bashkir)
- [đ§đž mGPT - 1.3B Belorussian](https://huggingface.co/ai - forever/mGPT - 1.3B - belorussian)
- [đ§đŦ mGPT - 1.3B Bulgarian](https://huggingface.co/ai - forever/mGPT - 1.3B - bulgarian)
- [đ mGPT - 1.3B Buryat](https://huggingface.co/ai - forever/mGPT - 1.3B - buryat)
- [đŗ mGPT - 1.3B Chuvash](https://huggingface.co/ai - forever/mGPT - 1.3B - chuvash)
- [đŦđĒ mGPT - 1.3B Georgian](https://huggingface.co/ai - forever/mGPT - 1.3B - georgian)
- [đ¸ mGPT - 1.3B Kalmyk](https://huggingface.co/ai - forever/mGPT - 1.3B - kalmyk)
- [đ°đŋ mGPT - 1.3B Kazakh](https://huggingface.co/ai - forever/mGPT - 1.3B - kazakh)
- [đ°đŦ mGPT - 1.3B Kirgiz](https://huggingface.co/ai - forever/mGPT - 1.3B - kirgiz)
- [đģ mGPT - 1.3B Mari](https://huggingface.co/ai - forever/mGPT - 1.3B - mari)
- [đ mGPT - 1.3B Ossetian](https://huggingface.co/ai - forever/mGPT - 1.3B - ossetian)
- [đŽđˇ mGPT - 1.3B Persian](https://huggingface.co/ai - forever/mGPT - 1.3B - persian)
- [đˇđ´ mGPT - 1.3B Romanian](https://huggingface.co/ai - forever/mGPT - 1.3B - romanian)
- [đšđ¯ mGPT - 1.3B Tajik](https://huggingface.co/ai - forever/mGPT - 1.3B - tajik)
- [â mGPT - 1.3B Tatar](https://huggingface.co/ai - forever/mGPT - 1.3B - tatar)
- [đšđ˛ mGPT - 1.3B Turkmen](https://huggingface.co/ai - forever/mGPT - 1.3B - turkmen)
- [đ mGPT - 1.3B Tuvan](https://huggingface.co/ai - forever/mGPT - 1.3B - tuvan)
- [đēđĻ mGPT - 1.3B Ukranian](https://huggingface.co/ai - forever/mGPT - 1.3B - ukranian)
- [đēđŋ mGPT - 1.3B Uzbek](https://huggingface.co/ai - forever/mGPT - 1.3B - uzbek)
- [đ mGPT - 1.3B Yakut](https://huggingface.co/ai - forever/mGPT - 1.3B - yakut)
đ License
This model is released under the MIT license.
Feedback
If you find a bug or have additional data for training the model on your language, please provide us with feedback.
The model will be improved over time. Stay tuned!
đ Information Table
Property |
Details |
Model Type |
Mongol mGPT 1.3B |
Training Data |
Additional data for 23 languages, base model trained on Wikipedia and C4 corpus for 61 languages from 25 language families |
License |
MIT |