Reinforcement learning from human feedback (RLHF)

Reinforcement learning, Reward model

Training large language models with reinforcement learning uses this method. See Bai2022training.

A nice explanation: https://gist.github.com/JoaoLages/c6f2dfd13d2484aa8bb0b2d567fbf093

An overview of the method

Models

Articles and resources