If I want to train with 4 GPUs, do I need to change the code
Opened this issue · 0 comments
alon2U commented
When I run the experiment using 4 GPUs, I will report an error:
RuntimeError: CUDA error: the launch timed out and was terminated
CUDA kernel errors might be asynchronously reported at some other API call,so the stacktrace below might be incorrect.