R

RLHF (Reinforcement Learning with Human Feedback)

ConceptMentioned in 1 video

A technique used to fine-tune large language models, questioned by LeCun for its efficiency and whether it's truly distinct from supervised learning.