position-sensitive attention
fmu2 opened this issue · 1 comments
fmu2 commented
Thanks for the great work!
I am a bit confused about this piece of code:
axial-deeplab/lib/models/axialnet.py
Line 67 in fe1d052
According to Eq. 4 in the paper, I have the impression that it should be torch.einsum('bgcj,cij->bgij', k, k_embedding) since p is the varying index. Please correct me if I am wrong. Thanks!
phj128 commented
This depends on the varying axis of the embedding you chooce, due to the two axis of the embedding here are two different directions, but both relative.