Get down and dirty with FlashAttention2.0 in pytorch, plug in and play no complex CUDA kernels
#3 opened 1 year ago in kyegomez/FlashAttention20
#2 opened 1 year ago in kyegomez/FlashAttention20