D

Deepseek R1 Quantized.w4a16

Developed by RedHatAI
INT4 weight-quantized version of DeepSeek-R1, reducing GPU memory and disk space requirements by approximately 50% while maintaining original model performance.
Downloads 119
Release Time : 4/17/2025

Model Overview

This model is a weight-quantized version of DeepSeek-R1, reducing weights from 8-bit to 4-bit, significantly lowering resource requirements while preserving the original model's performance. Suitable for large language model applications requiring efficient deployment.

Model Features

INT4 weight quantization
Reduces weights from 8-bit to 4-bit, decreasing GPU memory and disk space requirements by approximately 50%
Efficient deployment
Supports efficient deployment using vLLM backend, suitable for large-scale production environments
Performance retention
Maintains performance close to the original model after quantization

Model Capabilities

Text generation
Language understanding
Reasoning task processing

Use Cases

Education
Math problem solving
Solving complex math problems
Achieved 97.08% accuracy on MATH-500 test
Professional testing
AIME test
Handling American Invitational Mathematics Examination level problems
Achieved 77.00% accuracy on AIME 2024 test
General knowledge Q&A
MMLU test
Handling multidisciplinary multiple-choice questions
Achieved 86.99% accuracy on MMLU test
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase