RLHF (Reinforcement Learning with Human Feedback)
Concept
A technique used to fine-tune large language models, questioned by LeCun for its efficiency and whether it's truly distinct from supervised learning.
Mentioned in 1 video
A technique used to fine-tune large language models, questioned by LeCun for its efficiency and whether it's truly distinct from supervised learning.