M

Minicpm 2B Dpo Bf16

Developed by openbmb
MiniCPM is a series of edge large language models jointly open-sourced by Mianbi Intelligence and the Natural Language Processing Laboratory of Tsinghua University. The core language model, MiniCPM-2B, only contains 2.4 billion non-word embedding parameters.
Downloads 463
Release Time : 1/30/2024

Model Overview

MiniCPM-2B is a high-performance edge large language model. After supervised fine-tuning and direct preference optimization, it performs excellently in multiple evaluations, supports both Chinese and English, and can run on mobile phones through quantization.

Model Features

Efficient edge deployment
After Int4 quantization, it can achieve streaming inference on mobile phones, and the generation speed is slightly faster than human speech speed.
Multimodal capabilities
The edge multimodal large model MiniCPM-V built based on MiniCPM-2B has the best comprehensive performance among models of the same scale.
Low-cost secondary development
A single 1080/2080 graphics card can efficiently fine-tune parameters, and a 3090/4090 graphics card supports full-parameter fine-tuning. Continuous training can be completed on a single machine.
Excellent performance
In public comprehensive evaluations, it performs close to Mistral-7B, and its overall performance surpasses models such as Llama2-13B, MPT-30B, and Falcon-40B.

Model Capabilities

Text generation
Dialogue system
Code generation
Mathematical reasoning
Multimodal understanding

Use Cases

Education
Knowledge Q&A
Answer various subject knowledge questions
Performs excellently in Chinese, mathematics, and code ability evaluations
Development
Code assistance
Generate and optimize code
Performs well in code generation evaluations
Mobile applications
Mobile intelligent assistant
A language model application running on mobile phones
Supports streaming inference, and the generation speed is close to human speech speed
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase