gained 0.094 for eval_top1, and 0.33 for eval_top5, after 36-epoch training on 8 gpus
CheerM opened this issue · 4 comments
hi, would you mind releasing the training log for T2t-vit-t-14 training with 8 GPUs? I tried to rerun the script for training T2t-vit-t-14 with 8 GPUs. It gained 0.094 for eval_top1, 0.33 for eval_top5, after 36 epochs. It seems too slow to converge.
I don't know about the convergence over time but it should take 310 epochs to get the paper results.
hi, would you mind releasing the training log for T2t-vit-t-14 training with 8 GPUs? I tried to rerun the script for training T2t-vit-t-14 with 8 GPUs. It gained 0.094 for eval_top1, 0.33 for eval_top5, after 36 epochs. It seems too slow to converge.
Hi, the log of T2t-vit-t-14 is trained with 8 GPUs. It's normal if your results are slightly higher or lower than the logs.
hi, would you mind releasing the training log for T2t-vit-t-14 training with 8 GPUs? I tried to rerun the script for training T2t-vit-t-14 with 8 GPUs. It gained 0.094 for eval_top1, 0.33 for eval_top5, after 36 epochs. It seems too slow to converge.
Hello, have you solved the problem? I have the same problem. And the loss doesn't decrease.
Same here, training T2t-vit-t-14
on 3 GPU with -b64
, and after 80 epochs: top-1 acc = 0.095%, top-5 acc = 0.301%
It seems no improvements from epoch 20 (Top1=0.093, Top5=0.3199) through 80.