how many hours to train 1000 epochs
AugustusMJB opened this issue · 9 comments
hi,
with batchsize 128 and gtsrb's 50k images, how many hours do you used to train the 1000 epochs.
doese 1000 epochs necessary?
is the default setting ok for one gpu to train batchsize 128?
I use 8 gpus to train the batchsize 128, and 1st gpu has used 15g memory.
I am not sure what wrong with my reproduction for gtsrb2TT100k
hi, I have tried to use 1 gpu for training, it told me out of memory
Hi, I could not reproduce the memory error in my server.
What PyTorch version are you using?
The problem can be due to a different PyTorch version.
Hello,the "out of memory" problem is solved, I am not very clear about the reason, maybe it's a server problem. But the training speed is still slow, I have not found the reason yet. It takes 1.5 hour to train 8 epochs, which means I need 15 days for training 2000 epochs.
I will check it later, I will contact you if in need,
thank you very much