Edge 291: Reinforcement Learning with Human Feedback
1) Reinforcement Learning with Human Feedback(RLHF) 2) The RLHF paper, 3) The transformer reinforcement learning framework.
In this Issue:
Reinforcement Learning with Human Feedback(RLHF)
The RLHF paper.
The transformer reinforcement learning framework.
💡 ML Concept of the Day: Reinforcement Learnign with Human Feedback
One of the key improvements in models like ChatGPT or GPT-4 relative to its predecessors has been their ability to follow instructions. The genesis of this capability has its roots on a technique known as reinforcement learning with human feedback(RLHF) outlined in a 2017 paper. The core idea of RLHF is to extend LLM’s core feature of predicting the next word with the ability of understanding and fulfilling human requests. This is done by reformulating language tasks as a reinforcement learning problem.