B

Bert Base Arabic Camelbert Ca

Developed by CAMeL-Lab
CAMeLBERT is a collection of BERT models optimized for Arabic language variants, with the CA version specifically pre-trained on Classical Arabic texts
Downloads 1,128
Release Time : 3/2/2022

Model Overview

BERT model pre-trained on Classical Arabic (CA) datasets, suitable for fine-tuning Arabic NLP tasks

Model Features

Classical Arabic Optimization
Specifically pre-trained on 6GB of Classical Arabic text, excelling in CA tasks such as poetry classification (F1 80.9%)
Multi-task Adaptation
Supports 12 Arabic NLP tasks including NER, POS tagging, sentiment analysis, dialect identification, and poetry classification
Variant-Sensitive Processing
Preserves letter case and diacritics, employing whole-word masking strategy to enhance language feature learning

Model Capabilities

Masked language modeling
Next sentence prediction
Named entity recognition
Part-of-speech tagging
Sentiment analysis
Dialect identification
Poetry classification

Use Cases

Classical Literature Analysis
Arabic Poetry Classification
Automatic classification of Classical Arabic poetry
Achieved 80.9% F1 score on the APCD dataset
Linguistic Research
Classical Text Analysis
Analyzing linguistic features of Classical Arabic texts
Educational Technology
Arabic Learning Assistance
Helping learners understand Classical Arabic grammar and vocabulary
Featured Recommended AI Models
ยฉ 2025AIbase