R

Roberta Large Japanese

Developed by nlp-waseda
A large Japanese RoBERTa model pretrained on Japanese Wikipedia and the Japanese portion of CC-100, suitable for Japanese natural language processing tasks.
Downloads 227
Release Time : 5/10/2022

Model Overview

This is a large Japanese RoBERTa model pretrained on Japanese Wikipedia and the Japanese portion of CC-100, primarily used for masked language modeling and fine-tuning downstream tasks in Japanese text.

Model Features

Japanese-specific Pretraining
Specifically pretrained for Japanese text, optimizing performance for Japanese natural language processing tasks.
Juman++ Tokenization Support
Input text must be pre-tokenized using Juman++ to ensure efficient processing of Japanese text.
Large-scale Training Data
Trained on Japanese Wikipedia and the Japanese portion of CC-100, covering a wide range of Japanese corpus.
High-performance Hardware Training
Trained using eight NVIDIA A100 GPUs over two weeks to ensure high model quality.

Model Capabilities

Japanese Text Understanding
Masked Language Modeling
Downstream Task Fine-tuning

Use Cases

Natural Language Processing
Japanese Text Infilling
Use masked language modeling to fill in missing parts of Japanese text.
Downstream Task Fine-tuning
Fine-tune the model on specific Japanese NLP tasks such as text classification and named entity recognition.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase