LabeliaLabs/distributed-learning-contributivity

re-build the model reset the optimizer

arthurPignet opened this issue · 1 comments

When performing mpl, the model is erased and rebuilt every time we need it. Thus the model is recompiled, and the optimizer is reset, so the learning rate is not decreasing.

On the same subject: Adam or RMS opts use previous gradients in the descente direction. How does it work in fedavg ? For now on the optimizer is reset every time we need it, so this issue is hidden (we use this kind of gradient memory in the .fit() only, with gradient-pass-per-update batch). In fedavg, the weights are averaged, so how the optimizer reacts to this ?