GPT-2 paper

Study / Research

Referenced as the paper that introduced byte-level tokenization for large language models and motivated many tokenizer design decisions (vocab size, properties).

Mentioned in 1 video