R

Roberta Base Mr

Developed by flax-community
A transformers model pre-trained on large-scale Marathi corpus using self-supervised learning, primarily for masked language modeling and downstream task fine-tuning
Downloads 156
Release Time : 3/2/2022

Model Overview

Marathi model pre-trained with masked language modeling (MLM) objective based on RoBERTa architecture, suitable for sequence classification, token classification tasks

Model Features

Large-scale Marathi pre-training
Pre-trained on the mr subset of C4 multilingual dataset, containing 14 billion Marathi tokens
Dynamic masking mechanism
Unlike BERT, employs dynamic masking strategy during pre-training to enhance model generalization
Downstream task adaptability
Optimized for downstream tasks requiring whole-sentence understanding like sequence classification and token classification

Model Capabilities

Masked language modeling
Text classification
Sequence labeling

Use Cases

News classification
Marathi news headline classification
Classifying news headlines into 'state/entertainment/sports' categories
Test set accuracy 94.21%, outperforming iNLTK ULMFiT's 92.4%
IndicNLP news classification
Classifying news content into 'lifestyle/entertainment/sports' categories
Test set accuracy 97.48%, surpassing existing solutions
Featured Recommended AI Models
┬й 2025AIbase