# Chain-of-Thought Optimization

Fairyr1 32B GGUF
Apache-2.0
FairyR1-32B is an efficient large language model developed by Peking University DS-LAB, based on DeepSeek-R1-Distill-Qwen-32B. It achieves a balance between high performance and low-cost inference through an innovative 'distillation-fusion' process.
Large Language Model Transformers English
F
Mungert
867
1
Qwen3 30B A6B 16 Extreme 128k Context
A fine-tuned version of the Qwen3-30B-A3B mixture of experts model, with activated experts increased to 16 and context window expanded to 128k, suitable for complex reasoning scenarios
Large Language Model Transformers
Q
DavidAU
72
7
360zhinao3 7B O1.5
Apache-2.0
360 Zhinao 3-7B-O1.5 is a long chain-of-thought model open-sourced by Qihoo 360, fine-tuned based on 360 Zhinao 3-7B-Instruct, supporting complex reasoning tasks.
Large Language Model Transformers Supports Multiple Languages
3
qihoo360
35
3
MAI DS R1 FP8
MIT
MAI-DS-R1 is the result of Microsoft AI team's post-training of the DeepSeek-R1 inference model, aimed at enhancing its response capability to sensitive topics and optimizing risk performance, while maintaining the original reasoning ability and competitive advantages.
Large Language Model Transformers
M
microsoft
845
18
Deepseek R1 Medical CoT
MIT
A medical chain-of-thought reasoning model fine-tuned from DeepSeek-R1-Distill-Llama-8B, enhanced with QLoRA quantization and Unsloth optimization for medical reasoning.
Large Language Model
D
zijiechen156
49
2
Acemath 7B Instruct
AceMath-7B-Instruct is a specialized instruction model for mathematical reasoning developed by NVIDIA, based on an improved Qwen architecture, excelling at solving English math problems through chain-of-thought (CoT) reasoning.
Large Language Model English
A
nvidia
1,454
22
Llamav O1
Apache-2.0
LlamaV-o1 is an advanced multimodal large language model specifically designed for complex visual reasoning tasks, optimized through curriculum learning techniques, demonstrating outstanding performance across diverse benchmarks.
Text-to-Image Safetensors English
L
omkarthawakar
1,406
93
Codellama 7b Hf ReFT GSM8k
Enhances the reasoning generalization capabilities of large language models through reinforcement fine-tuning, based on Codellama fine-tuning, suitable for code generation and comprehension tasks.
Large Language Model Transformers
C
lqtrung1998
38
1
Llemma 7b
Llemma 34B is a language model specialized in mathematics, initialized with Code Llama 34B weights and trained on the Proof-Pile-2 dataset.
Large Language Model Transformers English
L
EleutherAI
3,668
106
Llama 30b Supercot
This model is a merge of huggyllama/llama-30b and kaiokendev/SuperCOT-LoRA, optimized for use with LangChain prompts, particularly suitable for logical reasoning and step-by-step thinking tasks.
Large Language Model Transformers
L
ausboss
115
125
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase