tgc1997/RMN

result reproduce for msr-vtt dataset

Closed this issue · 9 comments

Hi, Ganchao!
i have difficulty in reprodcing the experiment results for msr-vtt.
i have executed the project on msr-vtt several times and always got unideal results.
the cider scores just fluctuate from 45 to 46.5 which is far from the results i.e. 49.6 reported in the paper.
would it be convenient for u to share the random seed values set in ur experiments for msr-vtt with me?
training on msr-vtt is too time-consuming, 6 days or so when using a single gpu.
looking forward to ur help, thanks!

one of the training tensorboard logs is as follows:
2021-05-14 18-33-03 的屏幕截图
as illustrated in the picture, the light blue line represents one of the training records on msr-vtt dataset.
the output results :
BEST CIDEr(beam size = 2):
Bleu_1: 78.28
Bleu_2: 64.51
Bleu_3: 51.30
Bleu_4: 39.55
METEOR: 27.52
ROUGE_L: 59.84
CIDEr: 46.22
and the deep blue line is the training record that i executed the project again on msr-vtt without random seed settings yesterday
the later might be better than the former.

Hi! Does the latter one finishes training?

hi Ganchao!
thanks for ur attention to the training process.
the training is not over yet.
it will take 2 more days to finish all epochs.
so far, its logs is as follows:
2021-05-16 21-32-15 的屏幕截图
it seems like better than before.
could i stop training and evaluate it now? is it appropriate to do so?

hi, Ganchao!
the result of the latter experiment is as follows:
BEST CIDEr(beam size = 2):
Bleu_1: 79.03
Bleu_2: 65.13
Bleu_3: 51.72
Bleu_4: 40.05
METEOR: 27.88
ROUGE_L: 60.12
CIDEr: 46.59
BEST METEOR(beam size = 2):
Bleu_1: 79.63
Bleu_2: 65.68
Bleu_3: 52.00
Bleu_4: 40.06
METEOR: 28.07
ROUGE_L: 60.50
CIDEr: 47.36
and the logs screenshot is as follows:
2021-05-19 08-53-13 的屏幕截图
while the cider value is further improved in best-meteor epoch, there is still a gap to achieve the 49.6.
looking forward to your insights on the experiment, thanks!

What is the training batch size in your experiments? For msr-vtt, the best result I got is in the setting:--learning_rate=1e-4 --learning_rate_decay --learning_rate_decay_every=5 --learning_rate_decay_rate=3 --hidden_size=1300 --train_batch_size=48.
The results are as follows:(save 8 times for one epoch here)
7a96151573b82e7ad04cb5ba4bfb57e
BEST CIDEr(beam size = 2):
Bleu_1: 80.51
Bleu_2: 67.49
Bleu_3: 54.40
Bleu_4: 42.54
METEOR: 28.43
ROUGE_L: 61.62
CIDEr: 49.60

hi, Ganchao!
the training batch size set in all of my experiments is 8 due to gpu memory limitation.
so the key is the model performce will be affected by distinct batch size settings.
and would it be convenient for u to share the batch size settings for msvd dataset?
thanks for ur sincere sharing and generous help!

We set batch size to 32 for MSVD

hi, Ganchao~
i got it.
thanks again!