D

Deepseek R1 Distill Qwen 32B Unsloth Bnb 4bit

Developed by unsloth
DeepSeek-R1 is the first-generation inference model launched by the DeepSeek team. Through large-scale reinforcement learning training, it does not require supervised fine-tuning (SFT) as an initial step and demonstrates excellent inference capabilities.
Downloads 938
Release Time : 1/22/2025

Model Overview

The DeepSeek-R1 series of models focuses on inference tasks, has the ability of self-verification, reflection, and generating long chains of thought (CoT), and is suitable for mathematics, code, and inference tasks.

Model Features

Fast fine-tuning
Unsloth helps fine-tune large language models, increasing the speed by 2 - 5 times and reducing memory usage by 70%.
Powerful inference ability
Its performance in mathematics, code, and inference tasks is comparable to that of OpenAI-o1, and some distilled models outperform OpenAI-o1-mini.
Dynamic quantization
1.58-bit + 2-bit dynamic quantization, after selective quantization, significantly improves accuracy compared to standard 1-bit/2-bit quantization.
Open-source distilled models
Six dense models distilled from DeepSeek-R1 based on Llama and Qwen are open-sourced, providing more options for the research community.

Model Capabilities

Mathematical problem solving
Code generation
Long text inference
Self-verification
Reflection ability
Generating long chains of thought (CoT)

Use Cases

Mathematical problem solving
Solving math competition problems in AIME 2024
DeepSeek-R1 achieves a pass@1 of 79.8% on the AIME 2024 competition problems, surpassing GPT-4o and Claude-3.5-Sonnet.
79.8% pass@1
Solving 500 math problems in MATH
On the MATH-500 dataset, DeepSeek-R1 achieves a pass@1 of 97.3%, showing excellent performance.
97.3% pass@1
Code generation
Code generation on LiveCodeBench
DeepSeek-R1 achieves a pass@1-COT of 65.9% on LiveCodeBench, better than GPT-4o and Claude-3.5-Sonnet.
65.9% pass@1-COT
Solving programming competition problems on Codeforces
DeepSeek-R1 scores 2029 on the Codeforces competition problems, approaching the 2061 of OpenAI o1-1217.
2029 score
Inference tasks
Multi-task language understanding on MMLU
DeepSeek-R1 achieves a pass@1 of 90.8% on the MMLU dataset, showing excellent performance.
90.8% pass@1
Reading comprehension on DROP
On the DROP dataset, DeepSeek-R1's 3-shot F1 reaches 92.2%, surpassing GPT-4o and Claude-3.5-Sonnet.
92.2% 3-shot F1
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase