training parameters on Flickr30k dataset
mzhaoshuai opened this issue · 2 comments
mzhaoshuai commented
Hi, thx for your wonderful work!
I am trying to reproduce the Flickr30k captioning experiments in Table 1 in the paper, however, using the default parameters lr=2e-5
and epoch=10
, I get a much worse performance.
Could you please share your training parameters on Flickr30k dataset ?
DavidHuji commented
Hi, thanks for the interest in our work. I think on Flickr it requires more epochs, we trained until the validation results started to get worse. For COCO it was about 8 epochs but for smaller datasets (i.e. FlickrStyle10k), it was about 25 epochs.
mzhaoshuai commented
Thx for your response 😄 !