E

Elhberteu

Developed by orai-nlp
ElhBERTeu is a BERT model developed for the Basque language, trained on multi-domain corpora and demonstrating excellent performance in the BasqueGLUE benchmark.
Downloads 529
Release Time : 5/6/2022

Model Overview

ElhBERTeu is a base, case-sensitive monolingual BERT model for the Basque language, specifically designed for natural language understanding tasks, with a total of 124 million parameters.

Model Features

Multi-domain corpus training
Incorporates Basque texts from multiple domains including news, Wikipedia, science, and literature, totaling 575 million tokens.
Optimized training scheme
Completed 1 million pre-training steps on TPU with a sequence length of 512 and a batch_size of 256.
Outstanding benchmark performance
Achieved an average score of 73.71 in the BasqueGLUE benchmark, surpassing comparable models like BERTeus.

Model Capabilities

Basque text understanding
Named entity recognition
Intent classification
Slot filling
Text classification
Question answering systems
Word sense disambiguation
Coreference resolution

Use Cases

Natural Language Processing
Basque text classification
Automatic classification of Basque news, scientific literature, etc.
Achieved an F1 score of 78.05 in the BHTC task
Basque question answering system
Building intelligent question answering applications for Basque
Achieved an accuracy of 73.84 in the QNLI task
Linguistic research
Basque language analysis
Supports linguistic research on Basque grammar, semantics, etc.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase