Burmesebert
B
Burmesebert
Developed by jojo-ai-mst
Burmese-Bert is a bilingual masked language model based on bert-large-uncased, supporting both English and Burmese.
Large Language Model
Transformers Supports Multiple Languages#Burmese Language Understanding#Bilingual Masked Filling#BERT Architecture

Downloads 20
Release Time : 5/28/2024
Model Overview
This model is a bidirectional encoder representation based on the Transformer architecture, primarily used for Burmese natural language understanding tasks.
Model Features
Bilingual Support
Supports both Burmese and English processing
Based on BERT Architecture
Uses bert-large-uncased as the base model
Masked Language Modeling
Capable of predicting masked words in text
Model Capabilities
Burmese Text Understanding
English Text Understanding
Masked Word Prediction
Use Cases
Natural Language Processing
Burmese Text Completion
Automatically completes missing parts in Burmese text
Bilingual Text Analysis
Analyzes mixed texts containing Burmese and English
Featured Recommended AI Models
Š 2025AIbase