weiaicunzai/pytorch-cifar100

Problems about the architecture of Attention.

QJ-Chen opened this issue · 0 comments

In model.attention
AttentionModule1.shortcut_short is not used. You calculate the shortcut with the downsample weights.

shortcut_short = self.soft_resdown3(x_s)

AttentionModule3.shortcut_short is unnecessary.