seanie12/Info-HCVAE

About the Derivation of Variational Lower Bound in Appendix A

Hannofdhu opened this issue · 3 comments

I read the derivation you provided in appendix A, but I don't know how the last approximation is obtained, I hope to get your answer, thank you very much

It is very crude monte carlo approximation with sampling size 1 for z_x ~ q(z_x|x,c).

Thank you for your answer, I still don't understand something. Regarding the lower bound of the evidence derived in the paper, the coefficients in front of the KL divergence are all -1, why in the code, their coefficients become 0.1. Looking forward to your answer, thank you

Hi, sorry for the late reply.
If the coefficient is larger than -1 (beta=1 in the code), we cannot guarantee that it is lower bound of the marginal likelihood. But if we put the coefficient smaller than -1, we observe the posterior collapse. So, following [beta-vae],(https://openreview.net/forum?id=Sy2fzU9gl) we scale down the kl-divergence for training stability.