G

Gemma 3 12b It Quantized.w8a8

Developed by RedHatAI
An INT8 quantized version based on google/gemma-3-12b-it, supporting visual text input and text output, suitable for efficient inference deployment
Downloads 237
Release Time : 6/4/2025

Model Overview

This is a quantized multimodal model. It performs weight quantization on Gemma-3-12b-it and can be efficiently deployed with vLLM. It is suitable for scenarios with visual text input and text output

Model Features

Efficient quantization
Adopt INT8 weight quantization and INT8 activation quantization to significantly reduce model size and memory usage
Multimodal support
Support joint input of images and text for cross-modal understanding and generation
Efficient inference
Achieve efficient deployment through the vLLM backend, supporting batch processing and streaming output
High-precision maintenance
The quantized model maintains performance close to the original model in multiple benchmark tests

Model Capabilities

Image content understanding
Multimodal dialogue
Text generation
Visual question answering

Use Cases

Content understanding
Image description generation
Generate natural language descriptions based on input images
Can accurately describe the main content and scenes in the image
Visual question answering
Answer natural language questions about image content
Performs well in the MMMU and ChartQA benchmark tests
Intelligent assistant
Multimodal dialogue
Conduct natural conversations by combining image and text input
Can understand the image context and generate relevant responses
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase