M

Minicpm4 0.5B

Developed by openbmb
MiniCPM4 is an efficient large - language model designed specifically for edge devices. Through systematic innovation, it achieves extreme efficiency improvements in four key dimensions: model architecture, training data, training algorithm, and inference system.
Downloads 415
Release Time : 6/5/2025

Model Overview

The MiniCPM4 series is an efficient large - language model designed specifically for edge devices. The 0.5B version has 50 million parameters and is trained on 1T of tokens.

Model Features

Efficient model architecture
Adopting the trainable sparse attention mechanism of InfLLM v2, it significantly reduces the computational overhead when processing 128K long texts
Efficient learning algorithm
Including innovative technologies such as Model Wind Tunnel 2.0 and BitCPM extreme ternary quantization to achieve efficient training and compression
High - quality training data
Using UltraClean data filtering and generation technology to build high - quality pre - training and supervised fine - tuning datasets
Efficient inference system
Providing the CPM.cu lightweight CUDA inference framework and the ArkInfer cross - platform deployment system

Model Capabilities

Text generation
Dialogue interaction
Long text understanding
Tool invocation
Investigation report generation

Use Cases

Content creation
Article writing
Generate high - quality articles according to user prompts
The example shows the ability to generate AI - related articles
Tourism recommendation
Scenic spot recommendation
Recommend tourist attractions according to user needs
The example shows the recommendation of 5 tourist attractions in Beijing
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase