RLHF (Reinforcement Learning with Human Feedback)

Concept

A technique used to fine-tune large language models, questioned by LeCun for its efficiency and whether it's truly distinct from supervised learning.

Mentioned in 1 video