I

Indonesian Roberta Base

Developed by flax-community
Indonesian masked language model based on RoBERTa architecture, trained on the OSCAR corpus with a validation accuracy of 62.45%
Downloads 1,013
Release Time : 3/2/2022

Model Overview

This is a RoBERTa base model specifically optimized for Indonesian, suitable for various natural language processing tasks, especially masked language modeling.

Model Features

Indonesian-Specific Model
Pretrained model specifically optimized for Indonesian, excelling in Indonesian language tasks
Based on RoBERTa Architecture
Utilizes the proven RoBERTa architecture to provide robust language understanding capabilities
Efficient Training
Trained efficiently using Google Cloud's TPUv3-8 virtual machine, completed in just 18 hours and 25 minutes

Model Capabilities

Masked Language Modeling
Indonesian Text Understanding
Feature Extraction

Use Cases

Natural Language Processing
Text Completion
Predict masked words in sentences
Example: 'Budi sedang <mask> di sekolah.' can predict appropriate verbs
Feature Extraction
Extract text features for downstream NLP tasks
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase