re-build the model reset the optimizer
arthurPignet opened this issue · 1 comments
arthurPignet commented
When performing mpl, the model is erased and rebuilt every time we need it. Thus the model is recompiled, and the optimizer is reset, so the learning rate is not decreasing.
arthurPignet commented
On the same subject: Adam or RMS opts use previous gradients in the descente direction. How does it work in fedavg ? For now on the optimizer is reset every time we need it, so this issue is hidden (we use this kind of gradient memory in the .fit() only, with gradient-pass-per-update batch). In fedavg, the weights are averaged, so how the optimizer reacts to this ?