Gemma 3 4b It Qat 4bit
Gemma 3 4B IT QAT 4bit is a 4-bit quantized large language model trained with Quantization-Aware Training (QAT), based on the Gemma 3 architecture and optimized for the MLX framework.
Downloads 607
Release Time : 4/15/2025
Model Overview
This model is a quantized version converted from Hugging Face format to MLX format, suitable for image-text-to-text tasks with multilingual support.
Model Features
4-bit quantization
The model is trained with Quantization-Aware Training (QAT) to operate at 4-bit precision, significantly reducing memory usage and computational requirements.
MLX optimization
Optimized for the MLX framework, delivering efficient inference performance.
Multilingual support
Supports multiple languages, making it suitable for international applications.
Model Capabilities
Image-text generation
Multilingual text processing
Efficient inference
Use Cases
Image understanding and captioning
Image caption generation
Generates detailed textual descriptions based on input images.
Produces accurate and coherent image captions.
Multilingual applications
Multilingual image labeling
Generates multilingual labels or descriptions for images.
Supports image annotation in multiple languages.
Featured Recommended AI Models
Š 2025AIbase