R

Rumodernbert Base

Developed by deepvk
A modern bidirectional encoder-only Transformer model for Russian, pre-trained on approximately 2 trillion Russian, English, and code tokens, with a context length of up to 8,192 tokens.
Downloads 2,992
Release Time : 1/24/2025

Model Overview

RuModernBERT is the Russian version of ModernBERT, focusing on masked language model tasks and supporting both Russian and English.

Model Features

Long-context support
Supports context lengths of up to 8,192 tokens.
Multilingual pre-training
Pre-trained on Russian, English, and code tokens.
Efficient training
Utilizes a three-stage training strategy, including large-scale pre-training, context extension, and a cooling phase.

Model Capabilities

Masked language model
Russian text understanding
English text understanding

Use Cases

Natural Language Processing
Text completion
Predicts masked words or phrases.
Performs well in masked prediction tasks.
Text classification
Can be used for text classification tasks such as sentiment analysis.
Excels in the Russian Super Glue benchmark.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase