# Math & Code Enhancement
Llama 3.1 Nemotron Nano 4B V1.1
Other
Llama-3.1-Nemotron-Nano-4B-v1.1 is a large language model derived from Llama 3.1 8B through compression, optimized for inference efficiency and task execution, suitable for local deployment on a single RTX GPU.
Large Language Model
Transformers English

L
unsloth
219
4
Qwen3 32B GPTQ Int4
Apache-2.0
Qwen3 is the latest 8B parameter version of the Tongyi Qianwen series large language model, supporting thinking mode switching, multilingual processing, and tool invocation, with powerful reasoning and dialogue capabilities.
Large Language Model
Transformers

Q
JunHowie
1,079
3
Qwen3 14B Unsloth Bnb 4bit
Apache-2.0
Qwen3 is the latest generation of large language models in the Tongyi Qianwen series, offering both dense models and mixture-of-experts (MoE) models. Through large-scale training, Qwen3 achieves breakthrough progress in reasoning capabilities, instruction following, agent functionalities, and multilingual support.
Large Language Model
Transformers English

Q
unsloth
68.67k
5
GLM Z1 9B 0414
MIT
GLM-4-Z1-9B-0414 is the latest open-source model in the GLM family, featuring excellent mathematical reasoning and general capabilities, suitable for lightweight deployment in resource-constrained scenarios.
Large Language Model
Transformers Supports Multiple Languages

G
THUDM
3,456
55
Featured Recommended AI Models