# BPE tokenization

Birna Bert
A Transformer encoder model based on BERT architecture, specifically designed for generating RNA sequence embeddings
Text Embedding Transformers
B
buetnlpbio
364
1
Gena Lm Bert Large T2t
GENA-LM is an open-source foundational model family for long DNA sequences, based on a Transformer masked language model trained on human DNA sequences.
Molecular Model Transformers Other
G
AIRI-Institute
386
7
Deberta V1 Distill
Apache-2.0
A bidirectional encoder model pre-trained for Russian language, trained on large-scale text corpora using standard masked language modeling objectives
Large Language Model Transformers Supports Multiple Languages
D
deepvk
166
5
Dialogpt Medium Spanish Chitchat
This is a GPT-2 model based on the Transformer decoder, adjusted specifically for single-turn Spanish dialogue tasks, with 345 million parameters.
Dialogue System Transformers Spanish
D
ITG
1,890
4
Roberta Small Greek
This is a small Greek language model based on the RoBERTa architecture, with a parameter scale approximately half of the base model. It is suitable for the masked filling task of Greek text.
Large Language Model Transformers Other
R
ClassCat
22
2
Roberta TR Medium Bpe 44k
A RoBERTa model based on Turkish, pre-trained with masked language modeling (MLM) objective, case-insensitive.
Large Language Model Transformers Other
R
ctoraman
48
0
Roberta TR Medium Bpe 16k
A RoBERTa model pre-trained on Turkish with masked language modeling (MLM) objective, case-insensitive, medium-sized architecture.
Large Language Model Transformers Other
R
ctoraman
26
0
Rubert Base
Apache-2.0
Pre-trained Russian base BERT model by SberDevices team, with 178 million parameters, trained on 30GB Russian text
Large Language Model Transformers Other
R
ai-forever
29.43k
33
Kamo Naoyuki Mini An4 Asr Train Raw Bpe Valid.acc.best
This is an automatic speech recognition (ASR) pretrained model based on the ESPnet2 framework, trained on the mini-an4 dataset and supports English speech recognition.
Speech Recognition English
K
espnet
425
1
Rubert Large
A Russian large language model pre-trained by SberDevices team, based on Transformer architecture with 427 million parameters and 30GB training data
Large Language Model Transformers Other
R
ai-forever
6,125
14
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase