Behind the scenes: how Tri got started with Flash Attention
3
Motivation: modelling long sequences
4
Brief recap of attention
5
Memory bottleneck, IO awareness
6
Flash Attention 2.0 improvements
7
Behind the scenes of Flash Attention 2.0 refactor of CUTLASS 3
8
Future directions
9
Q&A
Description:
Dive into a comprehensive Discord server talk featuring Tri Dao from Stanford, discussing his groundbreaking work on Flash Attention 2.0. Explore the motivation behind modeling long sequences, gain insights into attention mechanisms, and understand the memory bottleneck and IO awareness challenges. Learn about the improvements in Flash Attention 2.0, including the refactoring of CUTLASS 3, and discover future directions in this field. Engage with an informative Q&A session to deepen your understanding of this cutting-edge technology in machine learning systems.
Flash Attention 2.0 with Tri Dao - Discord Server Talks