M

Medialbertina Pt Pt 900m

Developed by portugueseNLP
The first publicly available medical language model trained on real European Portuguese data
Downloads 70
Release Time : 3/12/2024

Model Overview

A Bert-series encoder based on DeBERTaV2 architecture, specifically optimized for the European Portuguese medical domain, obtained through continued pre-training of the Albertina model with electronic medical record data

Model Features

Medical Domain Specialization
Trained on 2.6 million electronic medical records from Portugal's largest public hospital, specifically optimized for the medical domain
European Portuguese Support
Focuses on European Portuguese variants, especially language usage in medical scenarios
Performance Improvement
Outperforms the base model in named entity recognition and assertion status classification tasks with significant F1 score improvements
Large-scale Training Data
Trained with over 15 million sentences and 300 million tokens of real medical data

Model Capabilities

Medical Text Understanding
Clinical Entity Recognition
Medical Terminology Filling
Electronic Medical Record Processing

Use Cases

Clinical Information Extraction
Symptom Recognition
Identify symptom descriptions from clinical texts
F1 score 0.832 (NER task)
Medication Dosage Recognition
Identify medication and dosage information in prescriptions
Can accurately fill medication dosage units
Medical Text Completion
Clinical Record Completion
Automatically complete incomplete clinical records
Can accurately predict medical terms as shown in examples
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Ā© 2025AIbase