Roberta Base Japanese
A Japanese RoBERTa-based pretrained model, trained on Japanese Wikipedia and the Japanese portion of CC-100.
Downloads 456
Release Time : 3/2/2022
Model Overview
This is a Japanese pretrained model based on the RoBERTa architecture, primarily used for masked language modeling tasks in Japanese. The model is trained on large-scale Japanese corpora and is suitable for various Japanese natural language processing tasks.
Model Features
Japanese-Specific Pretraining
Specifically pretrained for Japanese, using Japanese Wikipedia and the Japanese portion of CC-100 as training data
Juman++ Tokenization Support
Input text must be tokenized using Juman++ to ensure optimal processing of Japanese text
Large Vocabulary
Includes 32,000 tokens, combining JumanDIC vocabulary and subwords generated by sentencepiece
Efficient Training
Trained for one week using 8 NVIDIA A100 GPUs with various optimization techniques
Model Capabilities
Japanese Text Understanding
Masked Language Prediction
Downstream Task Fine-Tuning
Use Cases
Natural Language Processing
Text Completion
Predicts words replaced by the [MASK] token in sentences
Accurately predicts missing words in Japanese text
Text Classification
Can be fine-tuned for tasks like sentiment analysis and topic classification
Named Entity Recognition
Can be fine-tuned to identify entities such as person names and locations in Japanese text
Featured Recommended AI Models