when running s-mixup in DIG,we found the diffrernt results with the paper published inICML and the difference is large ,why?
ZDLTH opened this issue · 7 comments
What vale of alpha are you using?
As we shown in Table 7, S-Mixup is sensitive to alpha value.
when we let alpha equal to 1,we get 0.704 on IMDBB, when we let alpha equal to 0.5,we get 0.71 on IMDBB
What vale of alpha are you using? As we shown in Table 7, S-Mixup is sensitive to alpha value.
Hey Brother, can I add your WeChat to talk about the details of the operation of S-MixUp?
My WeChat ID is 18844094869
In my experiments, S-Mixup usually performs better when the alpha value is in the [0.1, 0.3] range. For IMDBB, I use alpha=0.1 and sim_function=abs_diff to get the best results. Please check Appendix B and C in our paper for hyperparameters. Let me know if there are any missing hyperparameters.
Please also make sure GMNET is well-trained, otherwise the results will be bad. If you don't want to train GMNET, you can try to use RRWM (https://pygmtools.readthedocs.io/en/latest/) to replace the computation of the soft alignments (https://github.com/divelab/DIG/blob/dig-stable/dig/auggraph/method/SMixup/smixup.py#L273).
Since ogbg-molhiv dataset has an official split and smaller variance, I also share my checkpoints of GMNET and GCN classifier at https://github.com/divelab/DIG_storage/tree/main/auggraph/SMixup. I hope this can help.
In my experiments, S-Mixup usually performs better when the alpha value is in the [0.1, 0.3] range. For IMDBB, I use alpha=0.1 and sim_function=abs_diff to get the best results. Please check Appendix B and C in our paper for hyperparameters. Let me know if there are any missing hyperparameters.
Please also make sure GMNET is well-trained, otherwise the results will be bad. If you don't want to train GMNET, you can try to use RRWM (https://pygmtools.readthedocs.io/en/latest/) to replace the computation of the soft alignments (https://github.com/divelab/DIG/blob/dig-stable/dig/auggraph/method/SMixup/smixup.py#L273).
Since ogbg-molhiv dataset has an official split and smaller variance, I also share my checkpoints of GMNET and GCN classifier at https://github.com/divelab/DIG_storage/tree/main/auggraph/SMixup. I hope this can help.
Thank you for your reply. My previous results were obtained using the optimal parameters provided in the appendix. Following your suggestion, I used alpha=0.1 and sim_function=abs_diff to run the IMDBB dataset, and the result was 0.7130
I just ran GIN on IMDB 3 times using the same hyperparameters and got 0.737, 0.719 and 0.729. It seems that our method is not stable. Thanks for bringing this up, I didn't notice this problem before. I think there are two potential reasons. One reason may be that the dataset is small, so the training of GMNET is not stable. Bad alignments computed by GMNET lead to changes in the results. Another reason could be the limitation we discussed in the paper, namely that the transformation of the graph makes the mixed label imperfect. When the dataset is small, imperfect labels for the mixed data can cause unstable results.