lucidrains/memory-efficient-attention-pytorch
Implementation of a memory efficient multi-head attention as proposed in the paper, "Self-attention Does Not Need O(n²) Memory"
PythonMIT
Stargazers
- achuthasubhashGUNTUR , INDIA
- anishthite@ClarosAI
- ben-zCanada
- brataoEscavador
- ChristophAltBayer
- ChristophReich1996Technical University of Munich
- dtch1997
- EmreTaha
- eugenesiowSingapore
- evdcush
- fly51flyPRIS
- hushell
- hysts
- JankinTian
- lkwq007CUHK
- lu-m13Intel Labs China
- mirzaskKansas City, MO
- monatis@qdrant
- mosh98Tietoevry
- mulkongNAVER Z
- nxznmNanjing University
- Question406UCSB
- sailfish009freelancer
- shyamsn97
- slyviacassell
- sriharsha0806fractal
- StillerPatrickHelmholtz AI
- styler00dollar
- theblackcat102iKala
- theerfanLos Angeles, CA
- vicgalleKomorebi AI & ICMAT-CSIC
- WelkinYangNetease FuxiAI
- xchanmolxManigos Information Technology Solutions
- XiSHEN0220Intellindust
- yuhangzangShanghai AI Laboratory
- zxczrx123Deeproute.ai