apache/mxnet

Can We implement Flash attention 2 in MXnet

rajveer43 opened this issue · 1 comments

Description

Flash Attention 2 is a library that provides attention operation kernels for faster and more memory efficient inference and training:

References

Welcome to Apache MXNet (incubating)! We are on a mission to democratize AI, and we are glad that you are contributing to it by opening this issue.
Please make sure to include all the relevant context, and one of the @apache/mxnet-committers will be here shortly.
If you are interested in contributing to our project, let us know! Also, be sure to check out our guide on contributing to MXNet and our development guides wiki.