Gemma 3 12B FornaxV.2 QAT CoT Q4 0 GGUF
This is an experimental small reasoning model designed to run on 8GiB consumer-grade GPUs with general inference capabilities. Through supervised fine-tuning (SFT) and high-quality reasoning trajectory training, the model can generalize its reasoning abilities to multiple tasks.
Downloads 98
Release Time : 5/6/2025
Model Overview
Gemma Fornax is a 12B-parameter model based on Gemma 3, focusing on general reasoning capabilities applicable to coding, mathematics, and other tasks. Trained with QAT checkpoints, the model maintains quality without degradation when used under Q4_0 quantization.
Model Features
General reasoning capability
Through supervised fine-tuning (SFT) with extensive and diverse high-quality reasoning trajectories, the model can generalize reasoning abilities to a wide range of tasks, not limited to programming and mathematics.
Low-resource requirement
The model is designed to run on 8GiB consumer-grade GPUs, requiring only about 6GiB memory under Q4_0 quantization.
Thinking mode switching
Similar to Qwen 3 series models, Gemma Fornax can enable or disable thinking modes via `/think` or `/no_think` instructions in system prompts.
QAT optimization
Training based on QAT checkpoints ensures no quality degradation when the model is used under Q4_0 quantization.
Model Capabilities
Text generation
General reasoning
Mathematical problem-solving
Programming assistance
Use Cases
Education and learning
Mathematical problem-solving
The model can solve complex mathematical problems and provide detailed reasoning processes.
Generates detailed problem-solving steps through thinking mode.
Programming development
Code generation and optimization
The model can generate code snippets or optimize existing code.
Produces efficient and executable code.
Creative writing
Story generation
The model can generate creative stories or character settings.
Creates coherent and imaginative text content.
Featured Recommended AI Models