G
GPT-2 paper
Study / ResearchMentioned in 1 video
Referenced as the paper that introduced byte-level tokenization for large language models and motivated many tokenizer design decisions (vocab size, properties).
Referenced as the paper that introduced byte-level tokenization for large language models and motivated many tokenizer design decisions (vocab size, properties).