D

Deepseek R1 GGUF

Developed by unsloth
DeepSeek-R1 is a 1.58-bit dynamically quantized large language model optimized by Unsloth, adopting the MoE architecture and supporting English task processing.
Downloads 2.0M
Release Time : 1/20/2025

Model Overview

Based on the quantized version of DeepSeek-R1, it achieves efficient inference through Unsloth's 1.58-bit + 2-bit dynamic quantization technology, and specifically optimizes the quantization strategy for the MoE layer.

Model Features

Dynamic hybrid quantization
Adopts 1.58-bit + 2-bit dynamic quantization technology to selectively quantize the MoE layer, with better precision than the standard 1-bit/2-bit scheme.
Multi-version quantization support
Provides four quantization schemes from 1.58bit to 2.51bit to balance disk usage and inference precision.
GPU acceleration optimization
Supports GPU layer offloading through llama.cpp to achieve accelerated inference on devices such as RTX 4090.

Model Capabilities

English text generation
Instruction following
Code understanding and generation

Use Cases

Development assistance
Code generation
Generate code in programming languages such as Python based on natural language descriptions.
The example demonstrates the ability to generate code for implementing the Flappy Bird game.
Content creation
Technical document translation
Handle mutual translation of Chinese and English technical documents.
Supports structured translation through specific prompt templates.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase