Can We implement Flash attention 2 in MXnet
rajveer43 opened this issue · 1 comments
rajveer43 commented
Description
Flash Attention 2 is a library that provides attention operation kernels for faster and more memory efficient inference and training:
References
github-actions commented
Welcome to Apache MXNet (incubating)! We are on a mission to democratize AI, and we are glad that you are contributing to it by opening this issue.
Please make sure to include all the relevant context, and one of the @apache/mxnet-committers will be here shortly.
If you are interested in contributing to our project, let us know! Also, be sure to check out our guide on contributing to MXNet and our development guides wiki.