# Precision adaptation

Qwen2.5 7B Instruct GGUF
Apache-2.0
Qwen2.5-7B-Instruct is an instruction-tuned model based on Qwen2.5-7B, optimized for text generation tasks, especially in chat scenarios.
Large Language Model English
Q
Mungert
706
4
Qwen2.5 3B Instruct GGUF
Other
Ultra-low-bit quantization (1-2 bit) model using IQ-DynamicGate technology, suitable for memory-constrained devices and efficient inference scenarios
Large Language Model English
Q
Mungert
704
4
Llama 3.2 3B Instruct GGUF
Llama-3.2-3B-Instruct GGUF is a 3B-parameter large language model released by Meta, utilizing IQ-DynamicGate technology for ultra-low-bit quantization (1-2 bits), optimizing inference performance while maintaining memory efficiency.
Large Language Model Supports Multiple Languages
L
Mungert
656
3
GLM Z1 32B 0414 GGUF
MIT
GLM-Z1-32B-0414 is a 32B-parameter multilingual text generation model supporting Chinese and English, released under the MIT license.
Large Language Model Supports Multiple Languages
G
Mungert
994
3
Deepcoder 14B Preview GGUF
MIT
Ultra-low-bit quantization (1-2 bits) model using IQ-DynamicGate technology, suitable for memory-constrained devices and edge computing scenarios
Large Language Model English
D
Mungert
1,764
6
Llama 3.1 Nemotron 70B Instruct HF GGUF
A model fine-tuned based on Meta Llama-3.1-70B-Instruct, optimized with NVIDIA HelpSteer2 dataset, supporting text generation tasks.
Large Language Model English
L
Mungert
1,434
3
Olmo 2 0325 32B Instruct GGUF
Apache-2.0
An instruction-tuned model based on OLMo-2-0325-32B-DPO, utilizing IQ-DynamicGate ultra-low bit quantization technology, optimized for memory-constrained environments.
Large Language Model English
O
Mungert
15.57k
2
Llama 3 3 Nemotron Super 49B V1 GGUF
Other
A 49B-parameter large language model utilizing IQ-DynamicGate ultra-low-bit quantization technology, supporting 1-2 bit precision-adaptive quantization, optimized for memory efficiency and inference speed
Large Language Model English
L
Mungert
434
4
Granite 3.2 8b Instruct GGUF
Apache-2.0
IBM Granite series 8B parameter instruction-tuned language model, employing IQ-DynamicGate ultra-low-bit quantization technology, suitable for efficient inference scenarios
Large Language Model
G
Mungert
1,048
2
Mistral Small 3.1 24B Instruct 2503 GGUF
Apache-2.0
This is an instruction-tuned model based on Mistral-Small-3.1-24B-Base-2503, utilizing GGUF format and IQ-DynamicGate ultra-low bit quantization technology.
Large Language Model Supports Multiple Languages
M
Mungert
10.01k
7
Qwen2.5 7B Instruct 1M GGUF
Apache-2.0
Qwen2.5-7B-Instruct-1M is an instruction-tuned version based on Qwen2.5-7B, employing IQ-DynamicGate ultra-low-bit quantization (1-2 bits), suitable for efficient inference in memory-constrained environments.
Large Language Model English
Q
Mungert
1,342
4
Qwen2.5 14B Instruct 1M GGUF
Apache-2.0
Qwen2.5-14B-Instruct-1M is an instruction-tuned model based on Qwen2.5-14B, supporting text generation tasks and suitable for chat scenarios.
Large Language Model English
Q
Mungert
1,600
3
Meta Llama 3 8B Instruct GGUF
An IQ-DynamicGate ultra-low-bit quantization (1-2 bit) model based on Llama-3-8B-Instruct, utilizing precision-adaptive quantization technology to enhance inference accuracy while maintaining extreme memory efficiency.
Large Language Model English
M
Mungert
1,343
3
Mistral 7B Instruct V0.1 GGUF
Apache-2.0
Mistral-7B-Instruct-v0.1 is a fine-tuned model based on Mistral-7B-v0.1, supporting text generation tasks. It employs IQ-DynamicGate ultra-low bit quantization technology, making it suitable for memory-constrained deployment environments.
Large Language Model
M
Mungert
632
3
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase