G

Gemma 3 12b It FP8 Dynamic

Developed by RedHatAI
An FP8 quantized model based on google/gemma-3-12b-it, supporting visual-text input and text output, suitable for multimodal scenarios.
Downloads 505
Release Time : 4/28/2025

Model Overview

This is a model that quantizes weights and activations using the FP8 data type. It can be efficiently deployed with vLLM and is suitable for scenarios with visual-text input and text output.

Model Features

FP8 Quantization
Both weights and activations are quantized using FP8 to improve inference efficiency.
Multimodal Support
Supports visual and text input and can handle prompts combining images and text.
Efficient Deployment
Can be efficiently deployed with the vLLM backend and supports services compatible with OpenAI.

Model Capabilities

Image Understanding
Text Generation
Multimodal Inference

Use Cases

Image Understanding and Description
Image Content Description
Generate descriptive text based on the input image.
Can accurately describe the image content.
Multimodal Q&A
Image-based Q&A
Answer questions about the image content.
Can generate relevant answers based on the image content.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase