M

Minicpm 1B Sft Bf16

Developed by openbmb
MiniCPM is an edge-side large language model series jointly developed by FaceWall Intelligence and Tsinghua University's Natural Language Processing Laboratory. The core model contains only 1.2 billion non-embedding parameters and outperforms larger open-source models in multiple evaluations.
Downloads 638
Release Time : 4/10/2024

Model Overview

MiniCPM is an efficient edge-side large language model series with excellent text generation and comprehension capabilities, supporting both Chinese and English, suitable for various natural language processing tasks.

Model Features

Efficient Performance
The 1.2B parameter model outperforms larger models like Llama2-13B and MPT-30B in multiple evaluations
Edge-side Deployment
After Int4 quantization, it can achieve streaming inference on mobile devices with generation speed approaching human speech rate
Multimodal Support
MiniCPM-V, built upon MiniCPM-2B, is the first multimodal large model that can run on mobile devices
Low-cost Development
Efficient fine-tuning can be done with a single 1080/2080 GPU, while 3090/4090 GPUs support full-parameter fine-tuning

Model Capabilities

Text generation
Dialogue systems
Instruction following
Mathematical reasoning
Code generation
Multimodal understanding

Use Cases

Intelligent Assistant
Knowledge Q&A
Answer various knowledge-based questions
Performance approaches Mistral-7B in public evaluations, with better Chinese/math/code capabilities
Education
Math Tutoring
Solve math problems and explain reasoning processes
Math capabilities surpass models of similar scale
Development Assistance
Code Generation
Generate code snippets based on descriptions
Code capabilities surpass models of similar scale
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase