B

Bert Base Uncased

Developed by deepvk
Pre-trained bidirectional encoder for Russian text processing, trained on large-scale social data and Wikipedia
Downloads 465
Release Time : 2/7/2023

Model Overview

Russian pre-trained model based on BERT architecture, specialized for text feature extraction tasks, excluding pre-trained head modules

Model Features

Russian optimization
Trained on 250GB Russian corpus including Wikipedia and social data
Pure encoder architecture
Contains only the encoder part, suitable for downstream task fine-tuning
Strict data filtering
Training data underwent rigorous quality control

Model Capabilities

Russian text feature extraction
Contextual semantic encoding
Downstream task fine-tuning foundation

Use Cases

Natural Language Processing
Text classification
Russian news classification/sentiment analysis
Achieved 0.467 accuracy on Russian Super Glue's RCB task
Question answering
Russian reading comprehension tasks
Achieved 0.737 accuracy on DaNetQA task
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase