G

Gemma 7b Zephyr Sft

Developed by wandb
A large language model based on Google's Gemma 7B, fine-tuned using the Zephyr SFT recipe, primarily for text generation tasks
Downloads 19
Release Time : 2/28/2024

Model Overview

This is an 8.5 billion-parameter GPT-style model, supervised fine-tuned (SFT) on a mix of public and synthetic datasets, focusing on English text generation tasks.

Model Features

Zephyr SFT Fine-tuning
Trained using the Zephyr supervised fine-tuning recipe, enhancing the model's performance in dialogue and instruction-following tasks
High-performance Benchmarking
Excels in multiple benchmarks on the Open LLM Leaderboard, such as HellaSwag (80.73) and Winogrande (74.19)
Computational Efficiency Optimization
Trained on 8xA100 80GB nodes, optimizing computational efficiency for large-scale models

Model Capabilities

Text Generation
Question Answering
Reasoning Tasks
Dialogue Systems
Knowledge QA

Use Cases

Education
AI2 Reasoning Challenge
Solving scientific reasoning problems
Standardized accuracy 61.43
Common Sense Reasoning
HellaSwag Evaluation
Common sense reasoning in everyday scenarios
Standardized accuracy 80.73
Winogrande Evaluation
Coreference resolution and common sense reasoning
Accuracy 74.19
Mathematical Ability
GSM8k Evaluation
Solving elementary school math problems
Accuracy 49.81
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase