L

Llm Jp 3 8x13b Instruct3

Developed by llm-jp
A large-scale Japanese-English hybrid MoE language model developed by Japan's National Institute of Informatics, supporting an 8x13B parameter scale with instruction fine-tuning optimization
Downloads 162
Release Time : 3/5/2025

Model Overview

A Transformer-based Mixture of Experts (MoE) model focused on Japanese and English text generation tasks, aligned through supervised fine-tuning and direct preference optimization

Model Features

Mixture of Experts Architecture
Utilizes an 8-expert routing design, activating only 2 experts per forward pass to balance computational efficiency and model capacity
Bilingual Optimization
Specially optimized for Japanese and English, with particularly outstanding performance in Japanese tasks
Safety Alignment
Optimized for preferences using safety datasets like AnswerCarefully, with a violation rate below 2.2%
Long Context Support
4096 tokens context window length, suitable for processing long documents

Model Capabilities

Japanese Text Generation
English Text Generation
Multi-turn Dialogue
Instruction Following
Knowledge Q&A
Text Summarization

Use Cases

Intelligent Assistant
Japanese Customer Service Bot
Handles inquiries and problem-solving for Japanese users
Achieved a reading comprehension score of 0.904 in llm-jp-eval tests
Educational Applications
Language Learning Assistance
Generates learning materials and exercises for Japanese and English
Featured Recommended AI Models
ยฉ 2025AIbase