using multi-GPU for training step
rafal7466 opened this issue · 1 comments
rafal7466 commented
Hi
Is there possible to use multi-GPU to decrease training time in your implementation?
gmberton commented
Hi,
we did not implement the multi-GPU training, because training the model is already very fast (around 1 day on a modern GPU). I think it should't be too difficult to do it with DistributedDataParallel, but we have no plans to do it.