Reinforcement learning from human feedback (RLHF)
Reinforcement learning, Reward model
Training large language models with reinforcement learning uses this method. See Bai2022training.
A nice explanation: https://gist.github.com/JoaoLages/c6f2dfd13d2484aa8bb0b2d567fbf093