FlashAttention-2: Making Transformers 800% faster AND exact

by kkenon 8/5/23, 6:11 PMwith 0 comments

This post has no comments