G

Gemma 3 12b It GPTQ 4b 128g

Developed by ISTA-DASLab
This model is an INT4 quantized version of google/gemma-3-12b-it, using the GPTQ algorithm to reduce parameters from 16-bit to 4-bit, significantly decreasing disk space and GPU memory requirements.
Downloads 1,175
Release Time : 4/11/2025

Model Overview

An INT4 quantized version based on Gemma-3-12b-it, suitable for text generation and multimodal tasks, maintaining most of the original model's performance while significantly reducing resource demands.

Model Features

Efficient INT4 Quantization
Uses the GPTQ algorithm to reduce parameters from 16-bit to 4-bit, cutting storage and memory requirements by approximately 75%.
Performance Retention
Maintains 98.42% of the original model's performance in OpenLLM benchmark tests.
Multimodal Support
Supports joint processing of images and text, capable of understanding and describing image content.

Model Capabilities

Text generation
Image content understanding
Multimodal task processing
Dialogue systems

Use Cases

Content generation
Image caption generation
Generates detailed descriptions based on input images
Accurately identifies objects and scenes in images and generates fluent descriptions
Intelligent assistant
Multimodal dialogue
Engages in dialogue combining image and text inputs
Understands image content and answers related questions
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase