Rubert Tiny
An extremely compact distilled version (45MB, 12M parameters) of the bert-base-multilingual-cased model for Russian and English, prioritizing speed and size over absolute accuracy
Downloads 36.18k
Release Time : 3/2/2022
Model Overview
This is a distilled miniature BERT model suitable for Russian and English tasks. The model is small, fast, and ideal for simple Russian tasks such as named entity recognition or sentiment classification. Its CLS embedding vectors can be used as sentence representations for Russian-English alignment.
Model Features
Compact Design
Approximately one-tenth the size and speed of standard BERT models, only 45MB in size
Bilingual Support
Supports both Russian and English processing, with embedding vectors enabling bilingual alignment
Multi-task Applicability
Suitable for various downstream tasks, including classification and named entity recognition
Efficient Distillation
Trained using MLM loss, translation ranking loss, and embedding vector distillation techniques
Model Capabilities
Fill-mask
Feature Extraction
Sentence Similarity Calculation
Text Classification
Named Entity Recognition
Use Cases
Natural Language Processing
Russian Text Classification
Performing sentiment analysis or topic classification on Russian texts
Bilingual Sentence Alignment
Using CLS embedding vectors for Russian-English sentence similarity calculation
Named Entity Recognition
Identifying entities such as person names and locations in Russian texts
Featured Recommended AI Models