R

Roberta Large Bne

Developed by PlanTL-GOB-ES
Large Spanish language model trained on data from the National Library of Spain (BNE)
Downloads 1,685
Release Time : 3/2/2022

Model Overview

Transformer-based Spanish masked language model using RoBERTa-large architecture, trained on 570GB of clean Spanish corpus

Model Features

Large-scale Spanish Pre-training
Trained on 570GB of clean, deduplicated text from the National Library of Spain (2009-2019)
Optimized RoBERTa Architecture
Utilizes RoBERTa-large architecture optimized for Spanish
Multi-task Applicability
Supports fine-tuning for various downstream NLP tasks such as Q&A, text classification, etc.

Model Capabilities

Masked Language Modeling
Text Feature Extraction
Named Entity Recognition
Text Classification
Question Answering System

Use Cases

Natural Language Processing
Text Classification
Classification tasks for Spanish text
F1 score of 0.9702 on MLDoc dataset
Named Entity Recognition
Identifying named entities in Spanish text
F1 score of 0.8823 on CoNLL-NERC dataset
Semantic Understanding
Semantic Similarity Calculation
Calculating semantic similarity between Spanish sentences
Composite score of 0.8411 on STS dataset
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase