Scroll

reinforcement learning from human feedback (RLHF)

A method in which AI agents learn from interactions with humans, receiving feedback on their performance to improve decision-making in a way that aligns with human preferences and goals.