D

Deepseek R1 0528 Quantized.w4a16

Developed by RedHatAI
The DeepSeek-R1-0528 model after quantization processing significantly reduces the requirements for GPU memory and disk space by quantizing the weights to the INT4 data type.
Downloads 126
Release Time : 5/30/2025

Model Overview

This model is a quantized version based on DeepSeek-R1-0528, mainly used for text generation tasks. Resource utilization efficiency is optimized through weight quantization.

Model Features

INT4 Weight Quantization
Reduce the weights from 8 bits to 4 bits, significantly reducing the GPU memory and disk space requirements by approximately 50%.
Efficient Deployment
Supports efficient deployment using the vLLM backend to optimize inference speed.
High-performance Inference
Performs close to the original model on multiple inference tasks with minimal accuracy loss.

Model Capabilities

Text Generation
Efficient Inference

Use Cases

Academic Research
Mathematical Problem Solving
Used to solve complex mathematical problems, such as the questions in the MATH-500 dataset.
pass@1 accuracy of 97.40%
General Knowledge Q&A
Answer the high-difficulty questions in the GPQA Diamond dataset.
pass@1 accuracy of 80.61%
Education
AIME Contest Question Answering
Generate answers to the questions in the American Invitational Mathematics Examination (AIME).
pass@1 accuracy of 87.33%
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase