R

Rhea 72b V0.5

Developed by davidkim205
Rhea-72b-v0.5 is a large language model fine-tuned based on Smaug-72B-v0.1, ranking first on the HuggingFace Open LLM Leaderboard.
Downloads 103
Release Time : 3/22/2024

Model Overview

The Rhea project researches various learning methods to improve large language model performance, using the nox framework for fine-tuning, combining supervised fine-tuning (SFT) and DPO learning methods.

Model Features

SGD method
Innovative self-generated dataset creation method for DPO learning, improving performance by comparing model outputs with correct answers.
High performance
Achieved a comprehensive score of 81.22 on the HuggingFace Open LLM Leaderboard, ranking first.
Diverse training data
Utilizes supervised fine-tuning dataset (datasets_enconv_4m) and DPO dataset (datasets_encomp_151k) from multiple sources.

Model Capabilities

Text generation
Reasoning ability
Q&A system
Mathematical computation
Language understanding

Use Cases

Academic research
AI2 Reasoning Challenge
Solving complex scientific reasoning problems
Standardized accuracy 79.78
MMLU test
Multidisciplinary knowledge understanding and application
Accuracy 77.95
Commercial applications
Math problem solving
Solving GSM8k math problems
Accuracy 76.12
Language understanding
HellaSwag commonsense reasoning
Standardized accuracy 91.15
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase