How do you conduct distributed training?
Polymorphy12 opened this issue · 1 comments
Polymorphy12 commented
Hello, I'm using 2 GPUs with a single node to train Atlas.
However, even if I set the local_rank to 0, the training doesn't start.
It still requires MASTER_ADDR, MASTER_PORT, etc.
Is there any additional information to notice?
mlomeli1 commented
I am afraid I can't help if you don't add more context of how exactly are you trying to do this.