Skip to content

Configuration for replicating results #535

@q8888620002

Description

@q8888620002

Hi

Thank you for developing and sharing this excellent library! I was wondering if it would be possible to provide the exact configurations used to replicate the results from the MolGPS paper, particularly the comparisons between MPNN++, GPS++, and Transformer.

It seems that some configurations in expts/neurips2023_configs may be outdated. Could you clarify if the latest configurations are available?

Also, I wonder if sharing an example yaml for multi-GPU training is possible. I've tried the following trainer configs

accelerator:
  type: gpu  # cpu or ipu or gpu
  config_override:
    datamodule:
      args:
        batch_size_training: 64
        batch_size_inference: 256
    trainer:
      trainer:
        precision: 32
trainer:
    precision: 32
    max_epochs: *max_epochs
    min_epochs: 1
    accumulate_grad_batches: 2
    check_val_every_n_epoch: 10
    devices: 4
    strategy: ddp

However, the following error shows up
RuntimeError: CUDA error: device-side assert triggered CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect.

Thanks in advance!

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions