Bert Tagalog Base Cased
A Tagalog BERT model trained on large-scale preprocessed text corpora to promote the development of the Filipino natural language processing community.
Downloads 74
Release Time : 3/2/2022
Model Overview
This is a BERT model specifically trained for Tagalog (Filipino), suitable for various natural language processing tasks.
Model Features
Large-scale corpus training
Trained on large-scale preprocessed text corpora collected and organized from the internet
Specialized for Tagalog
A BERT model specifically optimized for Tagalog, the primary language of the Philippines
Case-sensitive
The model is case-sensitive, suitable for application scenarios requiring case information retention
Model Capabilities
Text classification
Language understanding
Text feature extraction
Use Cases
Academic research
Low-resource language research
Used for studying natural language processing techniques for low-resource languages
Related research has been published on arXiv
Text processing
Tagalog text analysis
Used for processing and analyzing Tagalog text content
Featured Recommended AI Models
Š 2025AIbase