Roberta Large Japanese
A large Japanese RoBERTa model pretrained on Japanese Wikipedia and the Japanese portion of CC-100, suitable for Japanese natural language processing tasks.
Downloads 227
Release Time : 5/10/2022
Model Overview
This is a large Japanese RoBERTa model pretrained on Japanese Wikipedia and the Japanese portion of CC-100, primarily used for masked language modeling and fine-tuning downstream tasks in Japanese text.
Model Features
Japanese-specific Pretraining
Specifically pretrained for Japanese text, optimizing performance for Japanese natural language processing tasks.
Juman++ Tokenization Support
Input text must be pre-tokenized using Juman++ to ensure efficient processing of Japanese text.
Large-scale Training Data
Trained on Japanese Wikipedia and the Japanese portion of CC-100, covering a wide range of Japanese corpus.
High-performance Hardware Training
Trained using eight NVIDIA A100 GPUs over two weeks to ensure high model quality.
Model Capabilities
Japanese Text Understanding
Masked Language Modeling
Downstream Task Fine-tuning
Use Cases
Natural Language Processing
Japanese Text Infilling
Use masked language modeling to fill in missing parts of Japanese text.
Downstream Task Fine-tuning
Fine-tune the model on specific Japanese NLP tasks such as text classification and named entity recognition.
Featured Recommended AI Models