B

Bert Tagalog Base Cased

Developed by jcblaise
A Tagalog BERT model trained on large-scale preprocessed text corpora to promote the development of the Filipino natural language processing community.
Downloads 74
Release Time : 3/2/2022

Model Overview

This is a BERT model specifically trained for Tagalog (Filipino), suitable for various natural language processing tasks.

Model Features

Large-scale corpus training
Trained on large-scale preprocessed text corpora collected and organized from the internet
Specialized for Tagalog
A BERT model specifically optimized for Tagalog, the primary language of the Philippines
Case-sensitive
The model is case-sensitive, suitable for application scenarios requiring case information retention

Model Capabilities

Text classification
Language understanding
Text feature extraction

Use Cases

Academic research
Low-resource language research
Used for studying natural language processing techniques for low-resource languages
Related research has been published on arXiv
Text processing
Tagalog text analysis
Used for processing and analyzing Tagalog text content
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase