AIbase
Home
AI Tools
AI Models
MCP
AI NEWS
EN
Model Selection
Tags
Masked Language Prediction

# Masked Language Prediction

Pile T5 Base
Pile-T5 Base is an encoder-decoder model trained on The Pile dataset using the T5x library, trained for 2 million steps with MLM objective, approximately 2 trillion tokens.
Large Language Model Transformers English
P
EleutherAI
50
19
Serengeti E250
SERENGETI is a large-scale multilingual pre-trained model covering 517 African languages and dialects, focusing on bridging the gap in technological resources for African languages.
Large Language Model Transformers Other
S
UBC-NLP
42
6
Bert L12 H256 A4
A lightweight BERT model pretrained using knowledge distillation techniques, with a hidden layer dimension of 256 and 4 attention heads, suitable for masked language modeling tasks.
Large Language Model Transformers
B
eli4s
17
0
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
English简体中文繁體中文にほんご
© 2025AIbase