I

Icebert

Developed by mideind
Icelandic masked language model trained on RoBERTa-base architecture using 16GB of Icelandic text data
Downloads 1,203
Release Time : 3/2/2022

Model Overview

Pre-trained language model specifically designed for Icelandic, suitable for various natural language processing tasks

Model Features

Large-scale Icelandic training data
Integrated 7 different sources of Icelandic corpora, totaling 15.8GB of text data
Multi-domain coverage
Training data includes various text types such as news, medical literature, academic papers, and classical literature
Outstanding downstream task performance
Achieved state-of-the-art levels in tasks like part-of-speech tagging and named entity recognition

Model Capabilities

Text completion
Language understanding
Context prediction

Use Cases

Natural Language Processing
Part-of-speech tagging
Automatically identify the part-of-speech of words in Icelandic text
Achieved state-of-the-art performance
Named entity recognition
Identify entities such as person names and locations in Icelandic text
Achieved state-of-the-art performance
Text analysis
Grammar error detection
Detect grammatical errors in Icelandic text
Excellent performance
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase