D

Doge 320M

Developed by SmallDoge
Doge is a sequence transformation model that employs dynamic masked attention mechanisms, capable of state transitions using either multi-layer perceptrons or cross-domain mixture of experts.
Downloads 3,028
Release Time : 3/10/2025

Model Overview

The Doge model is trained by the SmallDoge community, supporting text generation tasks with dynamic masked attention mechanisms. It uses self-attention during training and state-space mechanisms during inference.

Model Features

Dynamic Masked Attention Mechanism
Allows the Transformer to use self-attention during training and state-space mechanisms during inference.
Cross-domain Mixture of Experts
Can directly inherit weights from multi-layer perceptrons for further training.
Efficient Training
Efficiently trained on RTX 4090 GPUs with relatively short training times.

Model Capabilities

Text Generation
Sequence Transformation

Use Cases

Natural Language Processing
Dialogue Generation
Can be used to generate natural language dialogue responses.
Generates fluent dialogue content
Content Creation
Can assist with writing and content creation.
Generates coherent text content
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase