training time is too long
Opened this issue · 1 comments
amsword commented
By default, 300 epochs are used for the training. On a machine with 4 P100, it needs about 21 days. Is it normal? How is the training time with V100 as described in Readme?
GOATmessi8 commented
@amsword About 6 days on 4 V100, with 64 batchsize. I have not tried to optimize many training parameters such as epochs and lr, and I think 200 epochs are enough to get a good result.