Reinforcement Learning Human Feedback (RLHF)

Concept

A post-training technique where AI refines its skills based on human feedback, akin to having a mentor or coach.

Mentioned in 1 video