U

URM LLaMa 3.1 8B

Developed by LxzGordon
URM-LLaMa-3.1-8B is an uncertainty-aware reward model designed to enhance the alignment of large language models.
Downloads 4,688
Release Time : 9/12/2024

Model Overview

This model consists of a base model and attribute-specific value heads with uncertainty awareness. It employs a two-stage training approach (attribute regression and gating layer learning) to provide more reliable reward signals.

Model Features

Uncertainty Awareness
The model can estimate the uncertainty of reward signals, with lower uncertainty indicating more reliable signals, leading to better alignment.
Two-stage Training
The first stage involves attribute regression training, while the second stage learns the gating layer to combine multi-attribute scores.
Gating Layer Learning
The gating layer dynamically combines multi-attribute scores through learning, rather than using fixed weights.

Model Capabilities

Text Quality Assessment
Reward Signal Generation
Uncertainty Estimation
Multi-attribute Scoring

Use Cases

Large Language Model Alignment
Response Quality Assessment
Evaluates the quality of AI assistant responses, including dimensions like helpfulness and correctness.
As shown in the charts, using uncertainty estimation leads to better alignment effects.
Reinforcement Learning
Reward Model
Provides more reliable reward signals for reinforcement learning training.
Reward signals with low uncertainty improve training stability.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase