The inconsistent performance on COLLAB dataset
ZhuYun97 opened this issue · 7 comments
Thanks for your fabulous codes, but I meet some problems while reproducing. After running 5 runs(sh scripts/run_graph.sh COLLAB 0
using seed: 0,1,2,3,4), the performance on COLLAB dataset is 78.78±0.46 which remains a large margin with the reported value(80.32±0.46).
I check the model configuration in the Appendix and find the hidden size of GIN is 512(but 256 in the config.yaml). After changing the hidden size to 512, the performance(78.70±0.32) is still inferior to the paper.
Is there anything I missing? hope for your reply.
@ZhuYun97 Thanks for your interest. The scaling coefficient alpha_l
should be set to 1 in COLLAB, while it is 2 by default now. Sorry to miss this hyper-parameter in configs.yml. We have updated configs.yml
. Hope this help.
Thanks for your instant reply.
Sorry to bother you again. The performance on MUTAG dataset is 86.47±1.81, but I find that all the hyperparameters are the same with Appendix. Is there any other things I should change?
The appendix shows only several key hyper-parameters due to space limits. You can refer to configs.yml
for complete information. And our experiments are conducted in NVIDIA 2080 Ti. Do you run the experiments using the hyper-parameters provided in configs.yml
? And maybe you could provide more details.
I use the latest config.yml
. And I run on NVIDIA 3090(I also test on NVIDIA 1080Ti, but get similar results).
MUTAG:
num_hidden: 32
num_layers: 5
lr: 0.0005
weight_decay: 0.00
mask_rate: 0.75
drop_edge_rate: 0.0
max_epoch: 20
encoder: gin
decoder: gin
activation: prelu
loss_fn: sce
scheduler: False
pooling: sum
batch_size: 64
alpha_l: 2
replace_rate: 0.1
norm: batchnorm
in_drop: 0.2
attn_drop: 0.1
alpha_l: 2
After I change the device to NIVIDIA 2080 Ti, the performance gets 88.29±1.22
. I guess the MUTAG dataset is too small, so the experiments are not so stable.
Indeed, MUTAG is too small, so the variance is larger and the results are not as stable as other results. And previously, we also found that some randomness can still exist in NIVIDIA 3090 Ti even when we use the same random seed. So you can just run the experiments on 2080 Ti.