S

Sroberta L

Developed by Andrija
RoBERTa language model trained on Croatian and Serbian, using 6GB dataset for 500,000 steps
Downloads 17
Release Time : 3/2/2022

Model Overview

This model is a Transformer-based language model optimized for Croatian and Serbian, primarily used for masked language modeling tasks.

Model Features

Multilingual support
Specifically optimized for Croatian and Serbian while also supporting multilingual processing
Large-scale training
Trained for 500,000 steps using 6GB of text data to ensure deep understanding of the target languages
RoBERTa architecture
Based on the third-generation RoBERTa architecture, providing advanced natural language processing capabilities

Model Capabilities

Text understanding
Language modeling
Masked prediction

Use Cases

Natural language processing
Text completion
Predict masked words in text
Language understanding
Analyze the semantics of Croatian and Serbian texts
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase