You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
First of all thank you for releasing this repository and providing the scripts to reproduce your paper,
it is deeply appreciated!
I have an issue reproducing the QSAR results from Table 3 in the paper for MolBERT and MolBERT (finetune),
as detailed below:
I can exactly reproduce the Table 3 entries for RDKit and ECFC4 using scripts/run_qsar_test_molbert.py so that is reassuring
The MolBERT featurizer, however, yields lower AUROCs i.e. for BACE I get 0.835 vs 0.849 from the paper and for BBBP I get versus 0.744 vs 0.750 in the paper.
Similarly for MolBERT (finetune) using scripts/run_finetuning.py for BBBP I get 0.751 vs 0.762 reported in the paper
Could it somehow be that I am using the wrong weights, or the wrong weights were uploaded to figshare? This would effect the results in both 2. & 3. above so would make sense.
Finally, the parameters I have been using for the fine-tuning are the following:
Hi MolBERT team,
First of all thank you for releasing this repository and providing the scripts to reproduce your paper,
it is deeply appreciated!
I have an issue reproducing the QSAR results from Table 3 in the paper for MolBERT and MolBERT (finetune),
as detailed below:
scripts/run_qsar_test_molbert.pyso that is reassuringscripts/run_finetuning.pyfor BBBP I get 0.751 vs 0.762 reported in the paperThe pre-trained model I am using is the one provided in the README i.e. https://ndownloader.figshare.com/files/25611290
Could it somehow be that I am using the wrong weights, or the wrong weights were uploaded to figshare? This would effect the results in both 2. & 3. above so would make sense.
Finally, the parameters I have been using for the fine-tuning are the following:
freeze_level = 0taken from the answer in Dataset size and creation #3learning_rate = 3e-5taken from the paper although I could only find the value for pre-training and not fine-tuningbatch_size=16All other arguments are left to the defaults provided in the code. Should the above arguments reproduce results similar to the paper?
Thanks in advance!
Tom