A

Albertina 100m Portuguese Ptbr Encoder

Developed by PORTULAN
Albertina 100M PTBR is a foundational large language model for Brazilian Portuguese, belonging to the BERT family of encoders, based on the Transformer neural network architecture, and developed on the DeBERTa model.
Downloads 131
Release Time : 5/25/2023

Model Overview

This model is a foundational large language model for Brazilian Portuguese with 100 million parameters, distributed under the MIT license, suitable for tasks such as masked language modeling.

Model Features

Optimized for Brazilian Portuguese
Specifically trained and optimized for Brazilian Portuguese, providing more accurate language understanding capabilities.
Based on DeBERTa Architecture
Developed on the DeBERTa model, combining the advantages of the Transformer architecture for stronger performance.
Open Source License
Distributed under the MIT license, allowing free use and modification.

Model Capabilities

Masked Language Modeling
Text Understanding
Downstream Task Fine-tuning

Use Cases

Natural Language Processing
Semantic Similarity Analysis
Can be used to analyze the semantic similarity between two texts.
Achieved a Pearson coefficient of 0.8501 on the STS-B task.
Textual Entailment Recognition
Determines whether one text entails the meaning of another.
Achieved an accuracy of 0.6462 on the RTE task.
Text Processing
Auto-completion
Predicts masked words.
Accurately predicted 'costumes' as the best completion word in the example.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase