# Inference Optimization
Nvidia AceReason Nemotron 7B GGUF
Other
AceReason-Nemotron-7B is a large language model based on the Nemotron architecture with 7B parameters, offering multiple quantized versions to accommodate different hardware requirements.
Large Language Model
N
bartowski
209
2
Nvidia AceReason Nemotron 14B GGUF
Other
AceReason-Nemotron-14B is a large language model with 14B parameters, offering multiple quantization versions to accommodate different hardware requirements.
Large Language Model
N
bartowski
1,772
6
Llama 3.1 Nemotron Nano 4B V1.1
Other
Llama-3.1-Nemotron-Nano-4B-v1.1 is a large language model derived from Llama 3.1 8B through compression, optimized for inference efficiency and task execution, suitable for local deployment on a single RTX GPU.
Large Language Model
Transformers English

L
unsloth
219
4
Mimo 7B RL
MIT
MiMo-7B-RL is a reinforcement learning model trained based on the MiMo-7B-SFT model, demonstrating outstanding performance in mathematical and code reasoning tasks, comparable to OpenAI o1-mini.
Large Language Model
Transformers

M
XiaomiMiMo
11.79k
252
Mimo 7B Base
MIT
A 7B-parameter specialized inference language model series launched by Xiaomi, significantly enhancing mathematical and code reasoning capabilities through optimized pre-training and post-training strategies
Large Language Model
Transformers

M
XiaomiMiMo
12.75k
101
Cognitivecomputations Dolphin3.0 R1 Mistral 24B GGUF
Dolphin3.0-R1-Mistral-24B is a 24B-parameter large language model based on the Mistral architecture, trained by Eric Hartford, focusing on reasoning and first-principles analysis.
Large Language Model English
C
bartowski
10.24k
72
Featured Recommended AI Models