R

Roberta Base Biomedical Es

Developed by PlanTL-GOB-ES
A RoBERTa architecture pretrained language model specifically designed for the Spanish biomedical domain, suitable for clinical text processing tasks
Downloads 335
Release Time : 3/2/2022

Model Overview

A medium-scale biomedical language model based on the RoBERTa architecture, trained on 963 million tokens of Spanish biomedical corpus, supporting masked language modeling tasks, particularly suitable for biomedical text named entity recognition and text classification

Model Features

Domain Specialization
Optimized specifically for the Spanish biomedical domain, outperforming general models in clinical text processing
High-Quality Corpus
Trained on a rigorously cleaned 963 million token biomedical corpus, integrating multiple authoritative medical data sources
Efficient Training
Using the same training parameters as the original RoBERTa, completed training in just 48 hours on 16 V100 GPUs

Model Capabilities

Biomedical Text Filling
Clinical Named Entity Recognition
Medical Text Classification
Biomedical Information Extraction

Use Cases

Clinical Document Processing
Medical History Auto-Completion
Automatically completes professional terminology in clinical records
Example shows 98.5% accuracy in completing 'arterial <mask>'
Radiology Report Analysis
Identifies abnormal descriptions in imaging reports
Accurately identifies abnormal descriptions in skeletal X-ray reports
Medical Research
Literature Information Extraction
Extracts key clinical information from medical literature
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase