Estbert
E
Estbert
Developed by tartuNLP
EstBERT is a BERT model pre-trained specifically for Estonian, supporting two sequence lengths of 128 and 512, and performing excellently in multiple natural language processing tasks.
Downloads 398
Release Time : 3/2/2022
Model Overview
EstBERT is a BERT model pre-trained on an Estonian corpus, mainly used for Estonian text understanding and processing tasks, such as part-of-speech tagging, named entity recognition, topic classification, and sentiment analysis.
Model Features
Dedicated to Estonian
Optimally trained specifically for Estonian, performing better than multilingual models on Estonian tasks.
Support for dual sequence lengths
Provides model versions with two sequence lengths of 128 and 512 to meet the needs of text processing of different lengths.
Superior comprehensive performance
Comprehensively outperforms mBERT and XLM-RoBERTa in tasks such as part-of-speech tagging, named entity recognition, topic classification, and sentiment analysis.
Model Capabilities
Part-of-speech tagging
Named entity recognition
Topic classification
Sentiment analysis
Text understanding
Masked language modeling
Use Cases
Natural language processing
Part-of-speech tagging
Perform part-of-speech tagging on Estonian texts
The UPOS accuracy reaches 97.89%, better than the comparison models
Sentiment analysis
Analyze the sentiment tendency of Estonian texts
The F1 score reaches 74.50, better than mBERT
Featured Recommended AI Models
Š 2025AIbase