apple/ml-mdm

[Question] Training another layer

Sys-if opened this issue · 1 comments

Hi,

First of all, thank you for sharing your code! I've been studying your paper and code, but I’m a bit unclear on how you implemented progressive training. Could you provide some insights into this? For example, do you freeze the inner U-Net for a certain number of iterations?

I'm replying to my own comment, as it might be useful for others. I have been experimenting with my own model and have implemented what I understood from the paper by progressively introducing the loss of the new high-resolution layer into the general loss, with a weight increasing from 0 to X. I also used a similar weight to progressively introduce the new layer features (as in Progressive Growing of GANs).