R

Rugpt3small Based On Gpt2

Developed by ai-forever
Russian pre-trained Transformer language model developed by SberDevices team, based on GPT2 architecture, supports 1024 sequence length, trained on 80 billion tokens.
Downloads 46.92k
Release Time : 3/2/2022

Model Overview

This model is a pre-trained language model optimized for Russian, designed based on the GPT2 architecture, trained on a large-scale Russian corpus, supporting long text generation and comprehension tasks.

Model Features

Russian optimization
Specially optimized for Russian language characteristics
Long sequence support
Supports 1024 sequence length, with fine-tuning extending to 2048 context window
Efficient training
Training completed in about one week on 32 GPUs

Model Capabilities

Russian text generation
Long text comprehension
Language modeling

Use Cases

Natural Language Processing
Russian text generation
Generate coherent Russian text content
Language model research
Serves as a baseline model for Russian NLP research
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase