Edit model card

GPT-2 Large RLHF Model for OpenAI TLDR Summarization

Based on SFT model ellipseai/gpt2-large-tldr-sum and perform RLHF training for better human alignment. The training curve on validation reward is

Validation reward during RL training

We perform evaluation for SFT model and RL model on 386 test set on the summarization quality with Claude-v2 to judge winner. We observe that RL model is significantly better than SFT model. This demonstrate that RL training is work very well.

model win loss tie win rate loss rate win rate adjusted
ellipseai/gpt2-large-tldr-sum 24 151 211 0.0622 0.3911 33.55%
ellipseai/gpt2-large-tldr-sum-rlhf 151 24 211 0.3911 0.0622 66.45%
Downloads last month
14
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.