You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I read the derivation you provided in appendix A, but I don't know how the last approximation is obtained, I hope to get your answer, thank you very much
The text was updated successfully, but these errors were encountered:
Thank you for your answer, I still don't understand something. Regarding the lower bound of the evidence derived in the paper, the coefficients in front of the KL divergence are all -1, why in the code, their coefficients become 0.1. Looking forward to your answer, thank you
Hi, sorry for the late reply.
If the coefficient is larger than -1 (beta=1 in the code), we cannot guarantee that it is lower bound of the marginal likelihood. But if we put the coefficient smaller than -1, we observe the posterior collapse. So, following [beta-vae],(https://openreview.net/forum?id=Sy2fzU9gl) we scale down the kl-divergence for training stability.
I read the derivation you provided in appendix A, but I don't know how the last approximation is obtained, I hope to get your answer, thank you very much
The text was updated successfully, but these errors were encountered: