# Chinese generation
Qwen3 8B 4bit DWQ
Apache-2.0
Qwen3-8B-4bit-DWQ is a 4-bit quantized version of Qwen/Qwen3-8B converted to the MLX format, optimized for efficient operation on Apple devices.
Large Language Model
Q
mlx-community
306
1
Qwen3 8B 4bit
Apache-2.0
This is the 4-bit quantized version of the Qwen/Qwen3-8B model, converted to the MLX framework format, suitable for efficient inference on Apple silicon devices.
Large Language Model
Q
mlx-community
2,131
2
Qwen3 4B 4bit
Apache-2.0
Qwen3-4B-4bit is a 4-bit quantized version converted from Qwen/Qwen3-4B to the MLX format, designed for efficient operation on Apple chips.
Large Language Model
Q
mlx-community
7,400
6
Qwq 32B Bnb 4bit
Apache-2.0
4-bit quantized version of QwQ-32B, optimized using Bitsandbytes technology, suitable for efficient inference in resource-constrained environments
Large Language Model
Transformers

Q
onekq-ai
167
2
Featured Recommended AI Models