Rugpt3large Based On Gpt2
Large-scale Russian pre-trained Transformer language model based on GPT-2 architecture, trained by SberDevices team
Downloads 9,985
Release Time : 3/2/2022
Model Overview
This model is a Transformer-based language model optimized for Russian, built on the GPT-2 architecture and pre-trained on a large corpus, suitable for Russian text generation and comprehension tasks.
Model Features
Large-scale Russian Pre-training
Pre-trained on an 80B token Russian corpus, specifically optimized for Russian
Long Context Support
Supports sequence lengths of 1024 and 2048, suitable for processing long texts
Efficient Training
Trained efficiently using 128 GPUs, with a total training time of approximately 14 days
Model Capabilities
Russian text generation
Russian text comprehension
Long text processing
Use Cases
Text Generation
Russian Article Creation
Generate coherent Russian articles or stories
Language Understanding
Russian Text Analysis
Understand and analyze Russian text content
Featured Recommended AI Models