# Multi-domain pre-training

Ko Llama3 Luxia 8B
A Korean-optimized large language model developed by Saltlux AI Lab based on Meta Llama-3-8B, featuring an extended Korean tokenizer and pre-trained on 100GB of curated Korean data
Large Language Model Transformers Supports Multiple Languages
K
saltlux
2,127
78
Vietnamese Llama2 7b 120GB
Other
A Vietnamese-optimized large language model based on Llama-2-7B, enhanced through continual pre-training on 124GB of multi-domain Vietnamese and English data for improved language understanding
Large Language Model Transformers Supports Multiple Languages
V
bkai-foundation-models
65
35
Mpt 7b
Apache-2.0
MPT-7B is an open-source commercial large language model trained by MosaicML. It is pre-trained on 1 trillion tokens of English text and code, and uses an improved Transformer architecture to optimize training and inference efficiency.
Large Language Model Transformers Other
M
mosaicml
27.19k
1,168
Kaz Roberta Conversational
Apache-2.0
Kaz-RoBERTa is a transformers model pre-trained in a self-supervised manner on a large-scale Kazakh corpus, primarily designed for masked language modeling tasks.
Large Language Model Transformers Other
K
kz-transformers
18.03k
13
Elhberteu
ElhBERTeu is a BERT model developed for the Basque language, trained on multi-domain corpora and demonstrating excellent performance in the BasqueGLUE benchmark.
Large Language Model Transformers Other
E
orai-nlp
529
2
Distilbert Mlm 1000k
DistilBERT is a lightweight distilled version of BERT, retaining 97% of BERT's performance while being 40% smaller and 60% faster.
Large Language Model Transformers
D
vocab-transformers
26
0
Distilbert Mlm 250k
DistilBERT is a lightweight distilled version of BERT, retaining most of BERT's performance but with fewer parameters and faster inference speed.
Large Language Model Transformers
D
vocab-transformers
17
0
CPM Generate
MIT
CPM is a Transformer-based 2.6 billion-parameter Chinese pre-trained language model, trained on 100GB of Chinese corpus, supporting various natural language processing tasks.
Large Language Model Transformers Chinese
C
TsinghuaAI
622
42
Gpt2
Indonesian generation model pre-trained based on causal language modeling objectives, trained using the Flax framework
Large Language Model Other
G
indonesian-nlp
130
11
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase