D

Deepseek R1 0528 AWQ

Developed by adamo1139
The 4-bit AWQ quantized version of the DeepSeek-R1-0528 671B model, suitable for use on high-end GPU nodes
Downloads 161
Release Time : 5/31/2025

Model Overview

This is the 4-bit AWQ quantized version of the DeepSeek-R1-0528 671B model, suitable for use with vLLM and SGLang on GPU nodes such as 8xA100/8xH20/8xH100.

Model Features

High-performance inference
Performs excellently in various benchmark evaluations such as mathematics, programming, and general logic, with overall performance approaching leading models such as O3 and Gemini 2.5 Pro
Quantization optimization
Uses 4-bit AWQ quantization technology to reduce the memory requirements during model operation while maintaining high performance
Enhanced inference depth
Significantly improves the inference depth and inference ability through the algorithm optimization mechanism in the later training
Reduced hallucination rate
Compared with the previous version, the new version reduces the hallucination rate and provides more accurate answers

Model Capabilities

Solving complex mathematical problems
Programming code generation
Logical reasoning
Multi-round dialogue
File content analysis
Integration of web search results

Use Cases

Education
Solving math competition problems
Solve math competition problems such as AIME and HMMT
Achieved an accuracy rate of 87.5% in the AIME 2025 test
Programming
Code generation and optimization
Generate and optimize programming code
Pass@1 reached 73.3% in the LiveCodeBench test
Research
Academic research assistance
Assist researchers in literature review and data analysis
Featured Recommended AI Models
ยฉ 2025AIbase