D

Devstral Small 2505.w4a16 Gptq

Developed by mratsim
This is a 4-bit GPTQ quantized version based on the mistralai/Devstral-Small-2505 model, optimized for consumer-grade hardware.
Downloads 557
Release Time : 5/25/2025

Model Overview

This model uses the asymmetric GPTQ method for 4-bit quantization (only 4-bit weights, W4A16) and is calibrated using 2048 samples with a maximum sequence length of 4096. It is suitable for text generation tasks.

Model Features

4-bit GPTQ quantization
The model is quantized to 4 bits (only 4-bit weights) using the asymmetric GPTQ method, significantly reducing hardware requirements
Optimized calibration strategy
Calibrated using 2048 samples with a maximum sequence length of 4096 to reduce the risk of overfitting and improve convergence
Consumer-grade hardware adaptation
Specifically optimized to run on consumer-grade GPUs (e.g., 32GB VRAM)

Model Capabilities

Text generation
Long sequence processing (up to 94000 tokens)

Use Cases

Code-related tasks
Code generation
Trained on the OpenCodeInstruct dataset, suitable for code generation tasks
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase