Gemma 7b Zephyr Sft
A large language model based on Google's Gemma 7B, fine-tuned using the Zephyr SFT recipe, primarily for text generation tasks
Downloads 19
Release Time : 2/28/2024
Model Overview
This is an 8.5 billion-parameter GPT-style model, supervised fine-tuned (SFT) on a mix of public and synthetic datasets, focusing on English text generation tasks.
Model Features
Zephyr SFT Fine-tuning
Trained using the Zephyr supervised fine-tuning recipe, enhancing the model's performance in dialogue and instruction-following tasks
High-performance Benchmarking
Excels in multiple benchmarks on the Open LLM Leaderboard, such as HellaSwag (80.73) and Winogrande (74.19)
Computational Efficiency Optimization
Trained on 8xA100 80GB nodes, optimizing computational efficiency for large-scale models
Model Capabilities
Text Generation
Question Answering
Reasoning Tasks
Dialogue Systems
Knowledge QA
Use Cases
Education
AI2 Reasoning Challenge
Solving scientific reasoning problems
Standardized accuracy 61.43
Common Sense Reasoning
HellaSwag Evaluation
Common sense reasoning in everyday scenarios
Standardized accuracy 80.73
Winogrande Evaluation
Coreference resolution and common sense reasoning
Accuracy 74.19
Mathematical Ability
GSM8k Evaluation
Solving elementary school math problems
Accuracy 49.81
Featured Recommended AI Models
Š 2025AIbase