R

Robbert V2 Dutch Base

Developed by pdelobelle
RobBERT is the current state-of-the-art Dutch BERT model, optimized based on the RoBERTa architecture, suitable for various text classification and tagging tasks
Downloads 7,891
Release Time : 3/2/2022

Model Overview

RobBERT is a large-scale pre-trained general Dutch language model that can be adapted to text classification, regression, or tagging tasks through fine-tuning. It achieves top performance in multiple Dutch NLP tasks, including sentiment analysis, coreference resolution, named entity recognition, etc.

Model Features

Dutch optimization
Specifically trained for Dutch, using a Dutch tokenizer and a large corpus of 6.6 billion words
RoBERTa architecture advantages
Adopts the optimized RoBERTa architecture, outperforming the original BERT
Few-shot learning capability
Performs exceptionally well in small dataset scenarios, significantly surpassing other models

Model Capabilities

Sentiment analysis
Coreference resolution
Named entity recognition
Part-of-speech tagging
Masked language modeling
Text classification

Use Cases

Sentiment analysis
Book review sentiment classification
Analyzing positive/negative sentiment in Dutch book reviews
95.1% accuracy, outperforming ULMFiT(93.8%) and BERTje(93.0%)
Syntax analysis
die/dat coreference resolution
Predicting whether to use 'die' or 'dat' in sentences
Full-data fine-tuning accuracy 99.23%, few-shot(10k) 97.82%
Part-of-speech tagging
Tagging parts of speech for Dutch text
96.4% accuracy on Lassy UD dataset, close to mBERT(96.5%)
Information extraction
Named entity recognition
Identifying entities like person names and locations in text
F1 score 89.08% on CoNLL 2002 dataset, close to mBERT(90.94%)
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase