wyhsirius/LIA

about architecture

Closed this issue · 1 comments

Hi, thanks for the nice work. I have some questions about architecture.

  1. In Fig.8 in, toRGB of shape 4 x 4, 512 not is used (refer to the red box in the below image), right? This is implemented by "self.to_rgb1=ToRGB(...)" in styledecoder.py, but didn't used in forward pass and also in backward pass. (This yields unused parameter error when training). I'm not sure this is intended implementation. I think this layer is important because it is the starting layer and the only the un-warped feature, which may helps to preserve id.

스크린샷 2022-09-25 오전 1 12 19

  1. Did you use NoiseInjection in Synthesis network? The parameters are assigned when initializing the NoiseInjection, but it is not used in the model.

If I missed something, please free to tell. Anyway, I am enjoying LIA. Thanks.

dd

i also found some other architecture difference not reported in the paper and i denote them on above figure.

  1. The attention mask "sigma" is also used as "1-sigma" for weight sum with the input. This is reasonable but not reported.
  2. There is additional skip layer which comes from the previous layer. As it comes the previous layer, it is upsampled x2 and added to the feature map which is output of conv 3x3 of ToFlow block. Can you explain this is right?

Thanks.