Bert Base Uncased
Pre-trained bidirectional encoder for Russian text processing, trained on large-scale social data and Wikipedia
Downloads 465
Release Time : 2/7/2023
Model Overview
Russian pre-trained model based on BERT architecture, specialized for text feature extraction tasks, excluding pre-trained head modules
Model Features
Russian optimization
Trained on 250GB Russian corpus including Wikipedia and social data
Pure encoder architecture
Contains only the encoder part, suitable for downstream task fine-tuning
Strict data filtering
Training data underwent rigorous quality control
Model Capabilities
Russian text feature extraction
Contextual semantic encoding
Downstream task fine-tuning foundation
Use Cases
Natural Language Processing
Text classification
Russian news classification/sentiment analysis
Achieved 0.467 accuracy on Russian Super Glue's RCB task
Question answering
Russian reading comprehension tasks
Achieved 0.737 accuracy on DaNetQA task
Featured Recommended AI Models