J

Japanese Gpt Neox 3.6b

Developed by rinna
A Japanese GPT-NeoX model with 3.6 billion parameters, based on the Transformer architecture, trained on 312.5 billion tokens of Japanese corpus.
Downloads 34.74k
Release Time : 5/17/2023

Model Overview

This is a Japanese language model based on the GPT-NeoX architecture, primarily used for text generation tasks, supporting Japanese natural language processing.

Model Features

Large-scale Japanese Pretraining
Trained on approximately 312.5 billion tokens of Japanese corpus, including CC-100, C4, and Japanese Wikipedia.
Optimized Tokenizer
Uses a sentencepiece-based tokenizer with UTF-8 byte fallback support, preserving whitespace information.
High Performance
Achieves a final validation perplexity of 8.68, indicating excellent language understanding capabilities.

Model Capabilities

Japanese Text Generation
Language Modeling
Natural Language Processing

Use Cases

Text Generation
Philosophical Text Continuation
Given the beginning of a philosophical topic, the model can generate coherent follow-up content.
Successfully generated coherent text about Nishida Kitaro's philosophy in the example.
Educational Research
Japanese Language Research
Can be used to study the performance and characteristics of Japanese language models.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase