S

Sinbert Large

Developed by NLPC-UOM
SinBERT is a Sinhala pre-trained language model based on the RoBERTa architecture, trained on a large Sinhala monolingual corpus (sin-cc-15M).
Downloads 150
Release Time : 3/2/2022

Model Overview

The SinBERT model is specifically designed for Sinhala text processing and is suitable for various natural language processing tasks, such as text classification.

Model Features

Sinhala optimization
Specifically pre-trained for Sinhala, optimizing text processing capabilities for this language.
Based on RoBERTa architecture
Adopts the RoBERTa architecture, inheriting its excellent natural language processing capabilities.
Large-scale training data
Trained using the sin-cc-15M large Sinhala monolingual corpus.

Model Capabilities

Text classification
Natural language understanding
Text feature extraction

Use Cases

Text analysis
Sinhala text classification
Performing classification tasks on Sinhala text
Demonstrated good classification performance in the LREC 2022 paper
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase