M

Minicpm 2B 128k

Developed by openbmb
MiniCPM is an edge-side large language model jointly developed by FaceWall Intelligence and Tsinghua University's Natural Language Processing Lab, with only 2.4 billion non-word embedding parameters (2.4B) and supports a 128k context window.
Downloads 145
Release Time : 4/9/2024

Model Overview

MiniCPM-2B-128k is an experimental long-context extension version based on MiniCPM-2B, being the first lightweight language model with less than 3B parameters that supports 128k context.

Model Features

128k long-context support
The first lightweight language model with less than 3B parameters that supports 128k context, achieving the best performance among sub-7B models in the comprehensive long-text benchmark InfiniteBench.
Lightweight model
With only 2.4 billion non-word embedding parameters (2.4B), it is suitable for edge-side deployment.
ChatML format support
Updated to ChatML format (user\n{}\nassistant\n) during model alignment, facilitating deployment via vLLM framework's OpenAI-compatible server mode.

Model Capabilities

Long-text processing
Dialogue generation
Text understanding
Knowledge Q&A

Use Cases

Long-text processing
Long-document Q&A
Handles document Q&A tasks with up to 128k context
Scored 23.06 in InfiniteBench's Chinese long-book Q&A evaluation
Dialogue systems
Intelligent assistant
Builds dialogue assistants with long-context memory support
Supports complex multi-turn dialogue interactions
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase