wayveai/mile

qusetions about training

Closed this issue · 0 comments

HITKJ commented

I attempted training on an 8xA800 machine with the following configuration:

N_WORKERS: 8
GPUS: 8
BATCHSIZE: 8
STEPS: 50000

# OPTIMIZER:
#   ACCUMULATE_GRAD_BATCHES: 1

RECEPTIVE_FIELD: 6
FUTURE_HORIZON: 6

VAL_CHECK_INTERVAL: 3000

image

When the iterations are proceeding normally, the CPU utilization appears as follows:
image

When there are no iterations happening, the CPU utilization looks like this:
image

My training time is close to 200 hours, but you mentioned that training with 8xV100 should only take two days. This has left me quite puzzled. My conclusion is that the current training bottleneck appears to be in the dataset processing part. When I set n_worker to 8, the progress bar gets stuck for a long time every 8 iterations, and GPU utilization drops to 0. Could there be any improper settings on my end? I look forward to your response.

(Note: I left the "OPTIMIZER" section commented out since it wasn't included in your original text.)