M

Melayubert

Developed by StevenLimcorn
A Malay masked language model based on the BERT architecture, trained on the Malay subset of the OSCAR dataset, supporting PyTorch and TensorFlow frameworks.
Downloads 15
Release Time : 3/2/2022

Model Overview

This model is a specialized masked language model for Malay, developed based on the BERT architecture, primarily used for masked prediction tasks in Malay text.

Model Features

Based on BERT Architecture
Utilizes the classic BERT architecture to ensure robust performance when processing Malay text.
Trained on OSCAR Dataset
The model is trained on the Malay subset of the OSCAR dataset, ensuring data diversity and comprehensiveness.
Supports PyTorch and TensorFlow
Compatible with two major deep learning frameworks, facilitating usage in different environments.

Model Capabilities

Masked Language Prediction
Malay Text Processing

Use Cases

Natural Language Processing
Malay Text Completion
Used to predict and complete masked portions in Malay text.
Achieved a perplexity score of 9.46 on the validation set.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase