K

Kunoichi DPO 7B

Developed by SanjiWatsuki
Kunoichi-DPO-7B is a model fine-tuned using Intel's Orca data for direct preference optimization (DPO) on the Alpaca template based on the Kunoichi-7B model. It is mainly targeted at general scenarios and has stronger reasoning and instruction-following abilities.
Downloads 748
Release Time : 1/11/2024

Model Overview

This model improves reasoning and instruction-following abilities through DPO fine-tuning and is suitable for general scenarios. However, the role-playing ability may be affected due to dataset alignment. It supports a maximum 8k context window and experimentally supports a 16k context window.

Model Features

Enhanced Reasoning Ability
Through DPO fine-tuning, the model performs better in reasoning and instruction following.
Large Context Window Support
Supports a maximum 8k context window and experimentally supports a 16k context window.
Suitable for General Scenarios
Suitable for various general scenarios, but the role-playing ability may decline.

Model Capabilities

Text Generation
Instruction Following
Reasoning Tasks

Use Cases

General Text Generation
Instruction Response
Generate appropriate response texts according to user instructions.
Scored 8.29 in the MT Bench test, outperforming multiple similar models.
Educational Assistance
Logic Test
Used for logical reasoning and problem-solving.
Scored 0.59 in the Logic Test, performing better than some similar models.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase