NormXU/ERNIE-Layout-Pytorch

Which version of Flash Attention has been used in this project?

Closed this issue · 2 comments

Hi,

I find the project is very interesting. Can I ask which Flash Attention has been used in this project?

From the official flash attention project, they have provided flash attention and flash attention v2.
https://github.com/Dao-AILab/flash-attention

Kind regards,
Qiming

@14H034160212 Thank you for your interest. I used PyTorch's Scaled Dot-Product Attention to speed up inference speed instead of Dao's implementations.

image
Do you use the FlashAttention-2 that implemented here? There are three supported implementations of scaled dot product attention.