A classic paper on Reinforcement Learning for Human Feedback (RLHF) is @[email protected]'s "Learning to summarize from human feedback".
Our talented engineer @[email protected] replicated this paper using our trlX library!
Read our report (w/ a code walk-through) here: http://wandb.me/summarize-rlhf-trlx