juntang-zhuang/Adabelief-Optimizer

loss become nan when beta1=0

yojeep opened this issue · 0 comments

Hello ,when I use Adabelief with beta1 = 0 , beta2 = 0.999(SAGAN、BIGGAN、WGAN-GP) ,the loss becomes nan , while Adam works well.I am wondering whether if the hyper parameter needs to be specifically changed when beta1 = 0? In many GANs, It is required that beta1 = 0 to stablize them.