Eleuther Pythia2.8b Hh Sft
A causal language model based on Pythia-2.8b, fine-tuned using the Anthropic human preference dataset
Downloads 205
Release Time : 8/7/2023
Model Overview
This is a large language model fine-tuned with Reinforcement Learning from Human Feedback (RLHF), focusing on generating text content that aligns with human preferences
Model Features
Human preference alignment
RLHF fine-tuning makes the model's outputs more aligned with human values and preferences
Transparency and reproducibility
Complete training logs and evaluation methods are publicly available to ensure research reproducibility
Efficient training
Significant performance improvement achieved with just 1 training epoch
Model Capabilities
Text generation
Dialogue systems
Preference-aligned text generation
Open-domain QA
Use Cases
Dialogue systems
Intelligent assistant
Building dialogue assistants that align with human preferences
Generates safer and more helpful responses
Content generation
Safe text generation
Generating content that adheres to ethical standards
Reduces the production of harmful or biased content
Featured Recommended AI Models