Hi
Thank you for developing and sharing this excellent library! I was wondering if it would be possible to provide the exact configurations used to replicate the results from the MolGPS paper, particularly the comparisons between MPNN++, GPS++, and Transformer.
It seems that some configurations in expts/neurips2023_configs may be outdated. Could you clarify if the latest configurations are available?
Also, I wonder if sharing an example yaml for multi-GPU training is possible. I've tried the following trainer configs
accelerator:
type: gpu # cpu or ipu or gpu
config_override:
datamodule:
args:
batch_size_training: 64
batch_size_inference: 256
trainer:
trainer:
precision: 32
trainer:
precision: 32
max_epochs: *max_epochs
min_epochs: 1
accumulate_grad_batches: 2
check_val_every_n_epoch: 10
devices: 4
strategy: ddp
However, the following error shows up
RuntimeError: CUDA error: device-side assert triggered CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect.
Thanks in advance!
Hi
Thank you for developing and sharing this excellent library! I was wondering if it would be possible to provide the exact configurations used to replicate the results from the MolGPS paper, particularly the comparisons between MPNN++, GPS++, and Transformer.
It seems that some configurations in
expts/neurips2023_configsmay be outdated. Could you clarify if the latest configurations are available?Also, I wonder if sharing an example yaml for multi-GPU training is possible. I've tried the following trainer configs
However, the following error shows up
RuntimeError: CUDA error: device-side assert triggered CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect.Thanks in advance!