AIbase
Home
AI Tools
AI Models
MCP
AI NEWS
EN
Model Selection
Tags
Efficient inference on the edge

# Efficient inference on the edge

Minicpm4 8B GGUF
Apache-2.0
MiniCPM4 is an efficient large language model designed specifically for edge devices. While maintaining optimal performance at the same scale, it achieves extreme efficiency improvements, enabling over 5x generation acceleration on typical edge chips.
Large Language Model Transformers Supports Multiple Languages
M
openbmb
324
1
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
English简体中文繁體中文にほんご
© 2025AIbase