G

Gemma 3 27b It FP8 Dynamic

Developed by RedHatAI
This is a quantized version of google/gemma-3-27b-it. The weights are quantized using the FP8 data type. It is suitable for visual-text input and text output, and can perform inference with efficient deployment using vLLM.
Downloads 1,608
Release Time : 4/28/2025

Model Overview

An FP8 quantized version based on the Gemma-3-27b-it model, supporting multimodal input (vision + text) and generating text output, with optimized inference efficiency.

Model Features

FP8 Quantization
Both weights and activations are quantized using the FP8 data type, significantly improving inference efficiency.
Multimodal Support
Supports joint input of vision and text, capable of understanding image content and generating relevant text.
Efficient Deployment
Optimized for the vLLM backend, supporting efficient inference deployment.
High Precision Maintenance
After quantization, the model maintains over 99% of the accuracy of the original model in multiple benchmark tests.

Model Capabilities

Image Content Understanding
Multimodal Text Generation
Visual Question Answering
Context Understanding

Use Cases

Image Understanding
Image Content Description
Generate a text description of the image content after inputting an image
Accurately identify the main elements and scenes in the image
Intelligent Question Answering
Visual Question Answering
Answer relevant questions based on the image content
Performs excellently in the MMMU and ChartQA benchmark tests
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase