fla-org/flash-linear-attention
🚀 Efficient implementations of state-of-the-art linear attention models in Torch and Triton
PythonMIT
Watchers
- 3outeilleHuggingFace
- apointa
- blaisI like to build beautiful things.
- cpuimageShantou, China
- dscamiss
- ghchris2021
- hannibalhuangShenzhen
- haojiweiBeihang University
- henry-tujiaInstitute of Computing Technology, Chinese Academy
- joshgong1977
- LegendBCHuazhong Uni. of Sci. and Tec.
- liyancThe University of Texas at Austin
- michalwolsNew York
- MonadKaiBeijing, China
- PengLU1101Canada
- renllMicrosoft
- Ryu1845
- sodabeta7Apple
- sourcesyncSan Francisco
- sustcsonglinMIT
- The-Dwh
- vapavlo
- vBaiCaiLi Auto
- wookayinUniversity of Michigan
- YangWang92Microsoft Research
- yzhangcsSoochow University
- zineos