H

Hh Rlhf Rm Open Llama 3b

Developed by weqweasdas
A reward model trained based on the LMFlow framework. It is trained on the HH - RLHF dataset (only the useful part) with open_llama_3b as the base model and has good generalization ability.
Downloads 483
Release Time : 7/14/2023

Model Overview

This reward model is used to evaluate the quality of conversation responses and can be used to generate reward signals in reinforcement learning. It supports multiple datasets.

Model Features

Strong Generalization Ability
The reward model trained on the HH - RLHF dataset performs excellently on the Open Assistant and chatbot datasets, even without direct training on these datasets.
Efficient Data Usage
By text splicing and splitting into blocks of size 1024 instead of padding according to the longest text, the data usage efficiency is improved.
High Accuracy
It achieves an accuracy of 75.48% on the HH - RLHF test set, with an evaluation loss of 0.5.

Model Capabilities

Conversation Response Quality Assessment
Reinforcement Learning Reward Signal Generation
Multi-dataset Generalization

Use Cases

Reinforcement Learning
RAFT Training
Use this reward model to generate high-quality responses in the RAFT framework for fine-tuning the GPT - Neo - 2.7B and LLaMA - 7B models.
The reward curve shows that the model can effectively improve the quality of generated responses.
Conversation System
Conversation Response Evaluation
Evaluate the quality of responses generated in the conversation system and select the optimal response.
It performs excellently on the Open Assistant and chatbot datasets.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase