M

Maltberta

Developed by MaCoCu
MaltBERTa is a large-scale pretrained language model based on Maltese text, using the RoBERTa architecture, developed by the MaCoCu project.
Downloads 26
Release Time : 8/11/2022

Model Overview

This model is specifically optimized for the Maltese language and is suitable for various natural language processing tasks.

Model Features

Large-scale Maltese pretraining
Trained on 3.2GB of Maltese text (439 million tokens)
Multi-source data integration
Combines content from MaCoCu, Oscar, and mc4 corpora, with deduplication processing
High-performance
Outperforms XLM-R-base/large on UPOS/XPOS and COPA benchmarks

Model Capabilities

Text understanding
Part-of-speech tagging
Language inference

Use Cases

Natural Language Processing
Part-of-speech tagging
Used for part-of-speech tagging tasks in Maltese text
Achieves 95.8/96.0 accuracy on the UPOS/XPOS test set
Language inference
Used for language inference tasks in Maltese
Achieves 53.7 accuracy on the COPA test set
Featured Recommended AI Models
ยฉ 2025AIbase