# Extreme quantization optimization
Minicpm4 8B GGUF
Apache-2.0
MiniCPM4-8B is an efficient large language model designed specifically for edge devices. Through innovations in four dimensions: model architecture, training data, training algorithms, and inference systems, it achieves extreme efficiency improvements.
Large Language Model
Transformers Supports Multiple Languages

M
Mungert
906
2
Gemma 3 1b It MAX HORROR Imatrix GGUF
Apache-2.0
A horror-optimized version based on Google's Gemma-3 model, featuring extreme quantization technology and horror enhancement matrix, supporting a 32k context window
Large Language Model
G
DavidAU
1,279
2
Gemma 3 12b It MAX HORROR Imatrix GGUF
Apache-2.0
A horror-style instruction-tuned version based on Google's Gemma-3 model, featuring Neo Imatrix technology and extreme quantization, supporting 128k context length
Large Language Model
G
DavidAU
5,072
13
Gemma 3 4b It MAX NEO Imatrix GGUF
Apache-2.0
An extreme quantization version based on Google's Gemma-3 model, enhanced with NEO Imatrix technology, supporting 128k context length and suitable for full-scenario tasks
Large Language Model
G
DavidAU
2,558
7
Featured Recommended AI Models