Roberta Base
A bidirectional encoder model pretrained specifically for Russian language, trained on large-scale text corpora, supporting feature extraction tasks.
Downloads 44
Release Time : 2/7/2023
Model Overview
This is a Russian pretrained model based on the RoBERTa architecture, primarily used for text feature extraction. The model was trained on a large-scale Russian corpus containing various text types such as social data, Wikipedia, and news articles.
Model Features
Russian Optimization
Specifically pretrained for Russian text, incorporating rich Russian linguistic features
Large-scale Training Data
Trained on 500GB of Russian text data, covering multiple text types and domains
High-performance Architecture
Based on RoBERTa architecture with 12 encoder layers and 768-dimensional embedding space
Model Capabilities
Russian text feature extraction
Contextual semantic understanding
Multidomain text processing
Use Cases
Natural Language Processing
Text Classification
Can be used for Russian text classification tasks
Performs well on the Russian Super Glue benchmark
Semantic Analysis
Suitable for semantic understanding and analysis of Russian text
Featured Recommended AI Models