Systematically different results from web server and local version
mapuyi opened this issue · 6 comments
I tried single runs from local installed version and also from the web server, and I got totally different behaviors of ligand poses even though I used exactly the same input files. I understand that every run should produce slightly different results, but the difference between local version and web server is obviously systematic that the local version gave ligand poses all around the protein while the web server gave most poses in the nearly accurate position.
I did not add any customized settings or "hack" into the backend script to change anything. I just followed the basic instructions and made it run. Is there any differences between web server and local version at the first place?
I also have the same issue.
Beside, I also have a problem: WARNING -Complex XXX Batch 1 Interfence Iteration XX smaple failed.
I also have the same issue. Beside, I also have a problem: WARNING -Complex XXX Batch 1 Interfence Iteration XX smaple failed.
That happens to me occasionally and I'm not sure why. Rerun several more times and this warning just doesn't show up anymore.
I have also noticed that in the docker container and web server I always meet this warning: parse_chi.py:91: RuntimeWarning: invalid value encountered in cast Y = indices.astype(int), but the results seem fine and most of the poses are in the nearly correct positions.
However in the local conda environment running, I never see this warning and always get terrible poses for the same input complex.
I have followed the default settings all the time. I also manually checked the scripts from docker container and did not find anything different.
I have also noticed that in the docker container and web server I always meet this warning: parse_chi.py:91: RuntimeWarning: invalid value encountered in cast Y = indices.astype(int), but the results seem fine and most of the poses are in the nearly correct positions.
However in the local conda environment running, I never see this warning and always get terrible poses for the same input complex.
I have followed the default settings all the time. I also manually checked the scripts from docker container and did not find anything different.
Thanks
Update here:
I re-installed all the packages under my local conda environment and this time I strictly followed the version numbers according to the authors' environment file (torch==1.13.1+cu117, torch_cluster===1.6.0+pt113cu117, and so on... actually all of them). Good news is now I can finally get similar results as Docker's and web-server's. It seems that some of those dependencies/packages have some sort of implicit impacts on the outputs.
@gcorso Could you guys check out those diffusion-model-related packages (torch, esm, and so on) with different versions/releases other than your environment requirement and see if there is really some impacts on the prediction outputs?
There's a reason we specify all of those version numbers. Informally I have observed the same as you, the results absolutely can change based on the library. I would generally expect the effect to be similar to picking a different random seed (assuming nothing breaks) but it isn't tested.