Skip to content

Missing HistoCLIP / HistoSigLIP code referenced in the paper #26

@duxiuju

Description

@duxiuju

Hello MarrLab team,

Thank you for releasing the HistoGPT codebase. I am trying to reproduce the experiments from the paper and noticed that the repository appears to include the main HistoGPT pipeline (MIL pre-training and autoregressive image-text fine-tuning), but I could not find the implementation/training scripts for the intermediate contrastive baselines HistoCLIP and HistoSigLIP described in the Methods.

In the paper you mention:

  • “For HistoCLIP we used the same loss as for CLIP. For HistoSigLIP we used the loss proposed in SigLIP.”
  • “We froze the vision encoder during training (locked-image text tuning).”

Could you please clarify:

  1. Are the HistoCLIP/HistoSigLIP training scripts available somewhere (another branch/repo), or are they planned to be released?
  2. If they are not planned for release, could you share the key implementation details needed for reproduction (e.g., exact image/text embeddings used, whether the resampler is trained or frozen, batch construction/negatives, temperature/logit_scale handling, and any loss weighting or normalization)?
  3. If possible, could you point to the commit/PR that contains these baselines?

Thanks in advance for your help.

Best regards,
Xiuju Du

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions