mlfoundations/wise-ft

Protocols for finetuning on CIFAR10 or CIFAR100.

alirezaabdollahpour opened this issue · 6 comments

Hi @gabrielilharco ;)
Could you kindly provide the protocols you use for finetuning on CIFAR10 or CIFAR100?
For example, factors such as the learning rate and the number of epochs... Please refer to Figure 16 in the main paper.

Hi @alirezaabdollahpour, the hyper-parameters are specified in Appendix D.2, please let us know if you have any questions

Hi @gabrielilharco , thank you for responding. In the section you mentioned earlier, you discussed the learning rates and augmentation. What is the recommended number of epochs for fine-tuning CIFAR10 models?

We do 10 epochs

I have a question about the warmup length in the paper.

The paper states that the warmup length is 500 steps, but for CIFAR100 with a batch size of 512, one epoch only takes 98 steps to train.
This means that the warmup phase occupies almost half of the total training time (total 10 epochs).
Is this correct?

It would be greatly helpful if you could clarify this point.
Thank you.

I believe so, do you find this drops accuracy compared to e.g. 50 warmup steps?