Tri Dao
Person
PhD graduate from Stanford, main author of the FlashAttention paper, and future Assistant Professor at Princeton. Currently Chief Scientist at Together.
Mentioned in 2 videos
Videos Mentioning Tri Dao

FlashAttention-2: Making Transformers 800% faster AND exact
Latent Space
PhD graduate from Stanford, main author of the FlashAttention paper, and future Assistant Professor at Princeton. Currently Chief Scientist at Together.

Stanford CS336 Language Modeling from Scratch | Spring 2026 | Lecture 4: Attention Alternatives
Stanford Online
One of the researchers credited with developing Mamba, Mamba 2, and Mamba 3.