Skip to content

Are we loading the whole dataset in the memory? #61

@joangog

Description

@joangog

Hi! I was trying to adapt the notebook of the downstream task of lung nodule segmentation (link) to brain tumor segmentation (BraTS). I noticed that the whole dataset is loaded in the memory and no pytorch dataloaders are used, nor custom dataloader classes are implemented. Is this the case? Because I am having trouble doing this for BraTS which is a large dataset. Could you perhaps provide the downstream task notebook for BraTS too, so that I can see how you managed to do this?

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions