Hi!
How do I get the model to utilise all available GPU memory on each GPU?
I tried changing --batch-len
, --optimizer
, --optimizer-opts
and some other parameters, but I can't seem to get it to use anything other than 416MiB per GPU.
Here I'm training 3 models in parallel:
Thanks!