Question about the KL Divergence
Closed this issue · 2 comments
Hi Tom,
Thanks for your great work.
Regarding your gradient descent for KL Divergence, I wonder is there any possibility that your optimized bsdfSamplingFraction will fall into a local optimization because the whole sphere domain share a same bsdfSamplingFraction?
Hi deadmarston,
the solution is definitely suboptimal because of the lack of dependence on incident direction as well as the averaging within the spatial SD-tree cell.
Alleviating either of these two limitations would make the learned BSDF sampling fraction more accurate.
Regarding local minima: applying SGD on the spatio-directionally averaged gradients amounts to finding an optimum of the spatio-directionally averaged loss function... so I believe you may be right about the averaging introducing local optima. Averaging multiple loss functions with a single global optimum can absolutely lead to multiple local optima, but I'm not able to make a definitive statement whether this'll happen in the specific use of the KL divergence that we have here (and under which conditions).
Cheers!
Hi Tom,
Thank you for your valueable reply:)
Cheers!