D

Doge 320M Instruct

Developed by SmallDoge
Doge 320M Instruct is a lightweight language model based on dynamic masked attention, trained with supervised fine-tuning (SFT) and direct preference optimization (DPO), suitable for question-answering and dialogue tasks.
Downloads 12.61k
Release Time : 3/25/2025

Model Overview

This model is trained by the SmallDoge community, using dynamic masked attention as sequence transformation and employing multi-layer perceptrons or cross-domain mixture of experts for state transformation. Suitable for English question-answering and dialogue tasks.

Model Features

Dynamic Masked Attention
Uses self-attention mechanism during training and state space during inference to improve efficiency
Cross-domain Mixture of Experts
Can directly inherit weights from multi-layer perceptrons for further training
Lightweight Design
Compact model with 320M parameters, suitable for resource-limited environments
Two-stage Training
First supervised fine-tuning (SFT), then direct preference optimization (DPO)

Model Capabilities

Text generation
Question-answering system
Dialogue system
Instruction following

Use Cases

Intelligent Assistant
Daily Conversation
Engages in natural language dialogue with users
Performs well on the SmolTalk dataset
Question-answering System
Knowledge Q&A
Answers various questions from users
Optimized on the UltraFeedback Binarized dataset
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase