T

Tanuki 8x8B Dpo V1.0

Developed by weblab-GENIAC
Tanuki-8x8B is a large-scale language model pretrained from scratch, optimized for dialogue tasks through SFT and DPO
Downloads 217
Release Time : 8/12/2024

Model Overview

Tanuki-8x8B-dpo-v1.0 is a large-scale language model with 8x8B parameters (total ~47B parameters, active ~13B parameters), pretrained on approximately 1.7T tokens, specifically optimized for Japanese and English dialogue tasks.

Model Features

Mixture of Experts architecture
Adopts an 8x8B Mixture of Experts architecture with ~47B total parameters but only ~13B active parameters, balancing performance and efficiency
Japanese optimization
Specifically optimized for Japanese dialogue tasks, supporting high-quality Japanese text generation
DPO optimization
Improved dialogue quality through Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO)
Multi-quantization support
Provides various quantization versions including AWQ and GPTQ for easy deployment in different hardware environments

Model Capabilities

Japanese text generation
English text generation
Multi-turn dialogue
Task-oriented dialogue

Use Cases

Intelligent assistant
Japanese Q&A system
Building intelligent Q&A assistants for Japanese users
Performed excellently in human evaluations
Education
Japanese learning assistance
Helping Japanese learners with language practice
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase