U

Ukr Roberta Base

Developed by youscan
RoBERTa base model trained on large-scale Ukrainian corpus, suitable for Ukrainian natural language processing tasks
Downloads 3,702
Release Time : 3/2/2022

Model Overview

This is a RoBERTa base model trained on Ukrainian Wikipedia, deduplicated OSCAR dataset, and social media texts, using the same architecture as roberta-base-cased and specifically optimized for Ukrainian

Model Features

Large-scale Ukrainian pre-training
Pre-trained on over 3.3 billion characters of Ukrainian corpus including Wikipedia, deduplicated OSCAR dataset, and social media texts
Standard RoBERTa architecture
Uses the same 12-layer Transformer architecture as roberta-base-cased, ensuring compatibility with other RoBERTa models
Efficient training
Trained in 85 hours using 4 V100 GPUs with high training efficiency

Model Capabilities

Text understanding
Text classification
Named entity recognition
Question answering
Text generation

Use Cases

Text analysis
Ukrainian sentiment analysis
Analyze sentiment tendencies in Ukrainian social media texts
Ukrainian news classification
Automatically classify Ukrainian news articles
Language understanding
Ukrainian question answering system
Build knowledge-based question answering systems for Ukrainian
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase