FlashAttention2 fa2.py contains 3 different attention implementations: Vanilla attention in PyTorch FlashAttention2 in PyTorch FlashAttention2 in Triton