R

Roberta TR Medium Morph 44k

Developed by ctoraman
A RoBERTa model for Turkish language, pre-trained with morphological-level tokenization and masked language modeling objectives, suitable for Turkish NLP tasks.
Downloads 453
Release Time : 3/9/2022

Model Overview

This model is a Turkish-optimized RoBERTa variant using morphological-level tokenization (with Zemberek morphological analyzer) and case-insensitive format, suitable for various Turkish text processing tasks.

Model Features

Morphological-Level Tokenization
Uses Zemberek Turkish morphological analyzer for text segmentation, achieving more Turkish-appropriate tokenization.
Case-Insensitive Format
Model input is case-insensitive, simplifying preprocessing and improving generalization.
Medium-Scale Architecture
Uses 8-layer Transformer structure for balanced computational efficiency and performance.

Model Capabilities

Turkish text understanding
Masked language modeling
Sequence classification (requires fine-tuning)

Use Cases

Natural Language Processing
Turkish Text Classification
Implement news classification, sentiment analysis etc. through model fine-tuning.
Language Model Pretraining
Serves as base model for transfer learning in Turkish NLP tasks.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase