FlashAttention 2: making Transformers 800% faster w/o approximation - with Tri Dao of Together AI - Latent Space: The AI Engineer Podcast - Podfriend