X

Xlm Roberta Base Finetuned Amharic Finetuned Ner Swahili

Developed by mbeukman
This is a token classification model (specifically for named entity recognition), which was fine-tuned on the Amharic language and then further fine-tuned on the Swahili portion of the MasakhaNER dataset.
Downloads 25
Release Time : 3/2/2022

Model Overview

This model is based on the Transformer architecture and was fine-tuned on the MasakhaNER dataset, primarily for named entity recognition tasks in Swahili.

Model Features

Cross-lingual Transfer Learning
The XLM-RoBERTa model, fine-tuned on Amharic, was further fine-tuned for Swahili, demonstrating the capability of cross-lingual transfer learning.
African Language Support
Optimized specifically for named entity recognition tasks in African languages (Swahili).
Efficient Training
Each model fine-tuning session took only 10-30 minutes, completed using an NVIDIA RTX3090 GPU.

Model Capabilities

Swahili Named Entity Recognition
News Domain Entity Extraction
Date, Location, Organization, and Person Name Recognition

Use Cases

News Analysis
News Entity Extraction
Extract key entity information from Swahili news articles
Overall F1 score of 86.66
Linguistic Research
African Language Processing
Used for natural language processing research on African languages
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase