X

XLMR MaCoCu Tr

Developed by MaCoCu
XLMR-MaCoCu-tr is a language model pretrained on large-scale Turkish text, part of the MaCoCu project, trained using 35GB of Turkish text.
Downloads 26
Release Time : 8/11/2022

Model Overview

This model is further trained based on the XLM-RoBERTa-large model, specifically designed for Turkish, suitable for various natural language processing tasks.

Model Features

Large-scale Turkish training
Trained with 35GB of Turkish text (4.4 billion tokens), covering a wide range of linguistic features.
Optimized based on XLM-RoBERTa-large
Further trained on XLM-RoBERTa-large, retaining the original vocabulary while optimizing Turkish language processing capabilities.
Superior multitask performance
Outperforms other Turkish language models in tasks such as POS tagging, NER, and COPA.

Model Capabilities

Part-of-speech tagging (UPOS/XPOS)
Named entity recognition (NER)
Causal reasoning (COPA)
Turkish text understanding

Use Cases

Natural Language Processing
Turkish text annotation
Used for part-of-speech tagging and named entity recognition in Turkish text.
Achieved 94.4% NER accuracy on the Universal Dependencies test set.
Causal reasoning
Used for Turkish causal reasoning tasks (COPA).
Achieved 60.7% accuracy on the MT test set, outperforming BERTurk and XLM-R-large.
Featured Recommended AI Models
ยฉ 2025AIbase