fla-org/flash-linear-attention
Efficient implementations of state-of-the-art linear attention models in Pytorch and Triton
PythonMIT
Watchers
- blaisNew York City
- sourcesyncSan Francisco
- hannibalhuangShenzhen
- wookayinNew York, NY
- michalwolsNew York
- YangWang92
- lcy-sesoChina
- vBaiCaiBeijing, China
- sodabeta7Cupertino
- renllRedmond, Washington
- apointa
- zineos
- yzhangcsShanghai
- dscamiss
- PengLU1101Canada
- drkostas
- realamirheTehran
- cpuimageShantou, China
- radarFudanSingapore
- liyancAustin, TX
- sustcsonglinCambridge
- LegendBCWuhan, China
- MonadKaiBeijing, China
- haojiwei
- 3outeilleFrance
- vapavlo
- ghchris2021
- Ryu1845
- joshgong1977