Roberta Small Japanese Aozora Char
A RoBERTa model pretrained on Aozora Bunko texts using a character tokenizer, suitable for Japanese text processing tasks.
Downloads 26
Release Time : 3/2/2022
Model Overview
This is a RoBERTa model pretrained on Aozora Bunko texts using a character tokenizer. It can be fine-tuned for downstream tasks such as part-of-speech tagging and dependency parsing.
Model Features
Character-level tokenization
Uses a character-level tokenizer, suitable for processing Japanese text
Aozora Bunko pretraining
Pretrained on Aozora Bunko texts, suitable for classical and modern Japanese text processing
Downstream task adaptation
Can be fine-tuned for various downstream NLP tasks such as part-of-speech tagging and dependency parsing
Model Capabilities
Masked language modeling
Japanese text understanding
Text feature extraction
Use Cases
Natural Language Processing
Part-of-speech tagging
Can be used for part-of-speech tagging tasks in Japanese text
Dependency parsing
Can be used to analyze the syntactic structure of Japanese text
Text completion
Can be used for masked prediction and completion of Japanese text
Featured Recommended AI Models
© 2025AIbase