R

Roberta Base

Developed by deepvk
A bidirectional encoder model pretrained specifically for Russian language, trained on large-scale text corpora, supporting feature extraction tasks.
Downloads 44
Release Time : 2/7/2023

Model Overview

This is a Russian pretrained model based on the RoBERTa architecture, primarily used for text feature extraction. The model was trained on a large-scale Russian corpus containing various text types such as social data, Wikipedia, and news articles.

Model Features

Russian Optimization
Specifically pretrained for Russian text, incorporating rich Russian linguistic features
Large-scale Training Data
Trained on 500GB of Russian text data, covering multiple text types and domains
High-performance Architecture
Based on RoBERTa architecture with 12 encoder layers and 768-dimensional embedding space

Model Capabilities

Russian text feature extraction
Contextual semantic understanding
Multidomain text processing

Use Cases

Natural Language Processing
Text Classification
Can be used for Russian text classification tasks
Performs well on the Russian Super Glue benchmark
Semantic Analysis
Suitable for semantic understanding and analysis of Russian text
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase