G

Gemma 2 9b It SimPO

Developed by princeton-nlp
Gemma 2.9B model fine-tuned on the gemma2-ultrafeedback-armorm dataset using the SimPO objective for preference optimization tasks
Downloads 21.34k
Release Time : 7/16/2024

Model Overview

This model is fine-tuned from Gemma 2.9B using the SimPO (Simple Preference Optimization) algorithm, designed to enhance large language model training through preference optimization datasets.

Model Features

SimPO Optimization Algorithm
Employs a simple preference optimization algorithm that requires no reference model, improving performance through reward function and generation probability alignment.
Efficient Training
Fine-tuning can be completed in approximately 100 minutes using just 8 H100 GPUs.
Performance Improvement
Outperforms the base model across multiple evaluation metrics.

Model Capabilities

Text Generation
Preference Optimization
Question Answering Systems
Dialogue Systems

Use Cases

Dialogue Systems
Intelligent Q&A
Used to build knowledge-based question answering systems
Achieved a score of 72.4 in AE2 LC evaluation
Content Generation
Long-Text Generation
Generates coherent long-form text
Average generation length of 1833 tokens
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase