Deepseek R1 0528 Quantized.w4a16
The DeepSeek-R1-0528 model after quantization processing significantly reduces the requirements for GPU memory and disk space by quantizing the weights to the INT4 data type.
Downloads 126
Release Time : 5/30/2025
Model Overview
This model is a quantized version based on DeepSeek-R1-0528, mainly used for text generation tasks. Resource utilization efficiency is optimized through weight quantization.
Model Features
INT4 Weight Quantization
Reduce the weights from 8 bits to 4 bits, significantly reducing the GPU memory and disk space requirements by approximately 50%.
Efficient Deployment
Supports efficient deployment using the vLLM backend to optimize inference speed.
High-performance Inference
Performs close to the original model on multiple inference tasks with minimal accuracy loss.
Model Capabilities
Text Generation
Efficient Inference
Use Cases
Academic Research
Mathematical Problem Solving
Used to solve complex mathematical problems, such as the questions in the MATH-500 dataset.
pass@1 accuracy of 97.40%
General Knowledge Q&A
Answer the high-difficulty questions in the GPQA Diamond dataset.
pass@1 accuracy of 80.61%
Education
AIME Contest Question Answering
Generate answers to the questions in the American Invitational Mathematics Examination (AIME).
pass@1 accuracy of 87.33%
Featured Recommended AI Models
Š 2025AIbase