B

Bert Medium Amharic

Developed by rasyosef
A pre-trained Amharic language model based on the bert-medium architecture, with 40.5 million parameters trained on 290 million tokens, achieving performance comparable to larger multilingual models.
Downloads 2,661
Release Time : 6/16/2024

Model Overview

A BERT model specifically designed for the Amharic language, supporting fill-mask tasks and applicable to text understanding and generation tasks.

Model Features

Efficient Parameter Utilization
Achieves Amharic language processing capabilities comparable to a 279 million parameter model with only 40.5 million parameters.
Dedicated Tokenizer
Amharic-specific tokenizer based on a 28k vocabulary.
Multi-dataset Training
Trained on integrated datasets including oscar, mc4, and Amharic sentence corpora.

Model Capabilities

Amharic text understanding
Fill-mask prediction
Downstream task fine-tuning

Use Cases

Natural Language Processing
Sentiment Analysis
Classify sentiment tendencies in Amharic text
F1 score 0.83
Named Entity Recognition
Identify entities such as person names and locations in Amharic text
F1 score 0.68
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase