G

Gemma 2B Rewardmodel Baseline

Developed by Ray2333
A scoring model based on Gemma-2b-it, trained with BT loss function, suitable for finding high-quality small scoring models for large language models
Downloads 133
Release Time : 7/5/2024

Model Overview

This model is a scoring model based on the Gemma-2b-it architecture, trained with BT loss function on the preference_700K dataset. It is primarily used to evaluate and select the output quality of large language models.

Model Features

Efficient scoring model
Small yet efficient scoring model, suitable for evaluating the output quality of large language models
BT loss function training
Optimized with Bradley-Terry (BT) loss function
Multi-dimensional evaluation capability
Capable of evaluating dialogue ability, safety, reasoning ability, and other dimensions

Model Capabilities

Text quality evaluation
Dialogue ability scoring
Safety evaluation
Reasoning ability scoring

Use Cases

Language model evaluation
LLM output quality evaluation
Evaluate the quality of text generated by large language models
Achieved a comprehensive score of 73.7 on the reward model benchmark
Dialogue system optimization
Used to optimize the response quality of dialogue systems
Dialogue ability score of 94.1
Content safety
Content safety filtering
Evaluate the safety of generated content
Safety score of 79.6
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase