M

Mistral Small 3.1 24B Instruct 2503 GPTQ 4b 128g

Developed by ISTA-DASLab
This model is an INT4 quantized version of Mistral-Small-3.1-24B-Instruct-2503, using the GPTQ algorithm to reduce weights from 16-bit to 4-bit, significantly decreasing disk size and GPU memory requirements.
Downloads 21.89k
Release Time : 3/20/2025

Model Overview

This model is a quantized version of Mistral-Small-3.1-24B-Instruct-2503, primarily designed for text generation tasks and supports multimodal input (image + text). It retains 97.8% of the original model's performance after quantization.

Model Features

Efficient Quantization
Utilizes INT4 quantization technology, reducing disk space and GPU memory requirements by 75%
High Performance Retention
Maintains 97.8% of the original model's performance after quantization
Multimodal Support
Supports joint input processing of images and text
Efficient Inference
The optimized model is suitable for deployment in resource-constrained environments

Model Capabilities

Text generation
Image caption generation
Multimodal understanding
Instruction following

Use Cases

Content generation
Image caption generation
Generate detailed descriptions based on input images
Can produce accurate and detailed image captions
Intelligent assistant
Multimodal dialogue
Process complex dialogues containing images and text
Capable of understanding and responding to complex queries involving visual information
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase