OpenNLPLab/cosFormer
[ICLR 2022] Official implementation of cosformer-attention in cosFormer: Rethinking Softmax in Attention
PythonApache-2.0
Issues
- 2
Attn Mask for Non-causal Models
#5 opened by roshansh-cmu - 3
Script for computing memory consumption
#12 opened by DaShenZi721 - 1
why input is [s b dim] but not [b s dim]?
#11 opened by Zyriix - 0
- 0
Question about space complexity
#8 opened by nihaomiao - 0
Pre-train model
#7 opened by csorujian - 1
- 1
- 1
Hello, when will the full code be published?
#3 opened by zmxsss - 1
When the code will be released?
#1 opened by LeeDoYup