B

Bert Base Arabert

Developed by aubmindlab
AraBERT is an Arabic pre-trained language model based on Google's BERT architecture, specifically designed for Arabic natural language understanding tasks.
Downloads 74.71k
Release Time : 3/2/2022

Model Overview

AraBERT is a BERT model optimized for Arabic, supporting various downstream NLP tasks, including sentiment analysis, named entity recognition, and question-answering systems.

Model Features

Arabic optimization
Pretrained specifically for Arabic characteristics, including handling prefix and suffix separation.
Farasa tokenization
Use the Farasa tokenizer for pre-segmentation to improve Arabic text processing.
Multi-task support
Support various downstream NLP tasks, including sentiment analysis, NER, and question-answering systems.
Large-scale pre-training
Pretrained using a 77GB Arabic corpus containing approximately 2.7B words.

Model Capabilities

Arabic text understanding
Sentiment analysis
Named entity recognition
Question-answering system
Text classification

Use Cases

Sentiment analysis
Arabic social media sentiment analysis
Analyze the sentiment of Arabic social media posts
Performs excellently on multiple Arabic sentiment analysis datasets such as HARD and ASTD
Information extraction
Arabic named entity recognition
Identify entities such as people's names and place names from Arabic text
Performs well on the ANERcorp dataset
Question-answering system
Arabic question-answering
Build an Arabic question-answering system
Has competitive performance on the Arabic-SQuAD and ARCD datasets
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase