B

Bert Large Mongolian Cased

Developed by tugstugi
This is a pre-trained Mongolian BERT model, trained on Mongolian Wikipedia and news datasets, supporting Mongolian text processing tasks.
Downloads 40
Release Time : 3/2/2022

Model Overview

This model is a Mongolian pre-trained model based on the BERT architecture, primarily used for masked language modeling tasks in Mongolian.

Model Features

Mongolian-specific
A pre-trained model specifically optimized for Mongolian text.
Case-sensitive
The model can recognize and handle case differences in Mongolian.
Large-scale training data
Trained on Mongolian Wikipedia and a news dataset with 700 million words.

Model Capabilities

Mongolian text understanding
Masked language modeling
Contextual semantic analysis

Use Cases

Text completion
Sentence completion
Automatically completes missing parts in Mongolian sentences.
As shown in examples, it can accurately predict words like 'нийслэл' (capital).
Language understanding
Semantic analysis
Understands the contextual meaning of Mongolian text.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase