D

Dots.llm1.inst

Developed by rednote-hilab
dots.llm1 is a large-scale MoE model that activates 14 billion parameters out of a total of 142 billion parameters, and its performance is comparable to that of the state-of-the-art models.
Downloads 440
Release Time : 5/14/2025

Model Overview

dots.llm1 is an open-source large-scale MoE model with an efficient data processing pipeline and high-performance inference capabilities, supporting both English and Chinese.

Model Features

Efficient Data Processing
Adopt a three-stage data processing framework to generate large-scale, high-quality, and diverse pre-training data.
Pretraining without Synthetic Data
The base model was pre-trained using 1.12 trillion high-quality non-synthetic tokens.
High Performance and Cost Efficiency
Only 14 billion parameters are activated during inference, combining comprehensive capabilities with high computational efficiency.
Innovative Infrastructure
Introduce an innovative MoE all-to-all communication and computation overlapping scheme based on interleaved 1F1B pipeline scheduling and efficient grouped GEMM implementation.
Open Model Dynamics
Intermediate model checkpoints trained every 1 trillion tokens were released to facilitate the study of the learning dynamics of large language models.

Model Capabilities

Text Generation
Dialogue System
Code Generation

Use Cases

Natural Language Processing
Text Completion
Used to generate coherent text completions, suitable for scenarios such as writing assistance and content generation.
Dialogue System
Used to build intelligent dialogue systems to provide a natural and smooth dialogue experience.
Programming Assistance
Code Generation
Used to generate code snippets, such as the implementation of the quicksort algorithm.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase