Reinforcement learning from human feedback: Revision history

Jump to navigation Jump to search

Diff selection: Mark the radio buttons of the revisions to compare and hit enter or the button at the bottom.
Legend: (cur) = difference with latest revision, (prev) = difference with preceding revision, m = minor edit.

28 April 2023

  • curprev 14:2514:25, 28 April 2023RobowaifuDev talk contribs 1,454 bytes +1,454 Created page with "'''Reinforcement learning from human feedback''' ('''RLHF''') is a subfield of reinforcement learning (RL) that trains agents using human feedback as reinforcement signals. In RL, agents interact with an environment, collect rewards or punishments based on actions taken and adjust their behavior to maximize rewards. However, designing accurate reward functions or annotating sufficient data for this purpose is difficult in many real-world scenarios. RLHF addresses th..."