Rugpt3small Based On Gpt2
Russian pre-trained Transformer language model developed by SberDevices team, based on GPT2 architecture, supports 1024 sequence length, trained on 80 billion tokens.
Downloads 46.92k
Release Time : 3/2/2022
Model Overview
This model is a pre-trained language model optimized for Russian, designed based on the GPT2 architecture, trained on a large-scale Russian corpus, supporting long text generation and comprehension tasks.
Model Features
Russian optimization
Specially optimized for Russian language characteristics
Long sequence support
Supports 1024 sequence length, with fine-tuning extending to 2048 context window
Efficient training
Training completed in about one week on 32 GPUs
Model Capabilities
Russian text generation
Long text comprehension
Language modeling
Use Cases
Natural Language Processing
Russian text generation
Generate coherent Russian text content
Language model research
Serves as a baseline model for Russian NLP research
Featured Recommended AI Models