M

Mistral Small 3.1 24B Instruct 2503 FP8 Dynamic

Developed by RedHatAI
This is a 24B-parameter conditional generation model based on the Mistral3 architecture, optimized with FP8 dynamic quantization, suitable for multilingual text generation and visual understanding tasks.
Downloads 2,650
Release Time : 3/27/2025

Model Overview

This model is the FP8 quantized version of Mistral-Small-3.1-24B-Instruct-2503, reducing GPU memory requirements and improving computational efficiency by lowering the bit-width of weights and activations while retaining most of the original model's performance.

Model Features

FP8 Quantization Optimization
Reduces GPU memory requirements and disk space by 50% through FP8 quantization while doubling matrix multiplication computational throughput.
Multilingual Support
Supports text generation and understanding in 24 languages.
Multimodal Capabilities
Capable of processing both image and text inputs for cross-modal understanding.
Efficient Inference
The optimized model is suitable for low-latency applications such as dialogue agents and function calling.

Model Capabilities

Text generation
Image understanding
Multilingual processing
Programming assistance
Mathematical reasoning
Long-document comprehension

Use Cases

Dialogue Systems
Intelligent Customer Service
Build fast-response multilingual dialogue agents
Maintains 97.9% of the original model's average performance in OpenLLM evaluation
Programming Assistance
Code Generation
Generate code based on natural language descriptions
Performs well in HumanEval evaluation
Education
Math Problem Solving
Solve mathematical reasoning problems
Achieves 49.96 points in GSM8K evaluation
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase