R

Rubert Tiny

Developed by cointegrated
An extremely compact distilled version (45MB, 12M parameters) of the bert-base-multilingual-cased model for Russian and English, prioritizing speed and size over absolute accuracy
Downloads 36.18k
Release Time : 3/2/2022

Model Overview

This is a distilled miniature BERT model suitable for Russian and English tasks. The model is small, fast, and ideal for simple Russian tasks such as named entity recognition or sentiment classification. Its CLS embedding vectors can be used as sentence representations for Russian-English alignment.

Model Features

Compact Design
Approximately one-tenth the size and speed of standard BERT models, only 45MB in size
Bilingual Support
Supports both Russian and English processing, with embedding vectors enabling bilingual alignment
Multi-task Applicability
Suitable for various downstream tasks, including classification and named entity recognition
Efficient Distillation
Trained using MLM loss, translation ranking loss, and embedding vector distillation techniques

Model Capabilities

Fill-mask
Feature Extraction
Sentence Similarity Calculation
Text Classification
Named Entity Recognition

Use Cases

Natural Language Processing
Russian Text Classification
Performing sentiment analysis or topic classification on Russian texts
Bilingual Sentence Alignment
Using CLS embedding vectors for Russian-English sentence similarity calculation
Named Entity Recognition
Identifying entities such as person names and locations in Russian texts
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase