tspeterkim/flash-attention-minimal
Flash Attention in ~100 lines of CUDA (forward pass only)
CudaApache-2.0
Stargazers
- Athe-kunalUnited States
- by2101
- clabrugere
- dmarxStability.ai, Eleuther.ai
- eclouder
- EdisonLeeeeeSun Yat-sen University
- FSSRepoFast Smart System
- HieuBui99Sun Asterisk
- JeffCarpenterCanada
- jinx2018SH
- ksanjeevanNew York
- KuangjuXUCAS
- laze44Xi an Jiaotong University
- lcy-sesoMSRA
- learning-chip
- Linaom1214Northwestern Polytechnical University
- lsj2408Peking University @microsoft
- MemorySlicesPrinceton University
- menegazzi
- mirceamironencoAmsterdam, Netherlands
- mitkotak
- okunatorUniversity of Helsinki
- RalphMao
- runrunrun1994
- Ryu1845
- SaltyPeppermintChair of Software Systems at Leipzig University
- shashankmangla
- shuuulPurdue University
- SushantDaga
- sustcsonglinMIT
- UlkuTuncerKucuktas
- walsvidByteDance
- wlruys
- XZK9Beijing
- YangWang92
- YdrMasterQiYuanLab