Japanese Gpt2 Medium
A medium-scale Japanese GPT-2 model trained by rinna Co., Ltd., based on the Transformer architecture, suitable for Japanese text generation tasks.
Large Language Model Supports Multiple LanguagesOpen Source License:MIT#Japanese Text Generation#24-layer Transformer#CC-100 Training
Downloads 7,664
Release Time : 3/2/2022
Model Overview
This is a Japanese language model based on the GPT-2 architecture, specifically optimized for Japanese text generation tasks.
Model Features
Japanese Optimization
Specifically trained and optimized for Japanese text
Medium Scale
Balances model performance and computational resource requirements
Transformer-based
Utilizes a 24-layer Transformer architecture with a 1024 hidden layer size
Model Capabilities
Japanese Text Generation
Language Modeling
Text Completion
Use Cases
Text Generation
Japanese Article Creation
Generate coherent Japanese articles or paragraphs
Dialogue Systems
Used for building Japanese chatbots
Language Processing
Text Completion
Complete Japanese sentences or paragraphs based on input prompts
Featured Recommended AI Models