Bert Medium Arabic
Pre-trained Arabic BERT medium language model, trained on approximately 8.2 billion words of Arabic text resources
Downloads 66
Release Time : 3/2/2022
Model Overview
This model is a pre-trained Arabic BERT medium language model, primarily used for natural language processing tasks such as text classification and named entity recognition.
Model Features
Multi-source Training Data
Trained on multiple Arabic text resources including OSCAR Arabic version and Wikipedia, totaling approximately 95GB.
Dialect Support
Includes not only Modern Standard Arabic but also some dialectal Arabic.
Optimized Training
Training parameter adjustments: Total steps of 3 million (batch size=128), instead of the original BERT's 1 million steps (batch size=256).
Model Capabilities
Text Classification
Named Entity Recognition
Text Generation
Language Understanding
Use Cases
Social Media Analysis
Offensive Speech Identification
Used to identify offensive speech on social media.
Performed well in SemEval-2020 Task 12.
Natural Language Processing
Text Classification
Used for Arabic text classification tasks.
Named Entity Recognition
Used to identify named entities in Arabic text.
Featured Recommended AI Models
Š 2025AIbase