Skip to content
/ DGMIL Public

Official PyTorch implementation of our MICCAI 2022 paper: DGMIL: Distribution Guided Multiple Instance Learning for Whole Slide Image Classification.

Notifications You must be signed in to change notification settings

miccaiif/DGMIL

Repository files navigation

🍄 DGMIL

This is a PyTorch/GPU implementation of our MICCAI 2022 paper DGMIL: Distribution Guided Multiple Instance Learning for Whole Slide Image Classification.

Main models and training frameworks are uploaded. For patch generating, please follow DSMIL for details. For MAE pretraining, please follow MAE for details.

Description of Key Inputs

MAE_dynamic_trainingneg_feats.npy

  • Array of dimensions n*512, where n represents the features of negative instances within the training set.
  • 512 is the dimensionality of features extracted by MAE.

MAE_dynamic_trainingpos_feats.npy

  • Features of positive instances within the training set.
  • Due to the absence of instance labels, positive instances here refer to all instances from positive slides.

MAE_testing_neg_feats.npy & MAE_testing_pos_feats.npy

  • Features of all negative and positive instances with true instance labels within the testing set, respectively.
  • These are used for testing and metric calculation. Both are in the format of n*512.

test_slide_label.npy & num_bag_list_index.npy

  • Used for slide-level prediction.
  • Implementation needs improvement. Currently, instances need to be assembled into a bag feature based on num_bag_list_index.
  • Alternatively, any method of reading features on a Slide basis can be considered. A better attempt can be referenced from here.

MAE_dynamic_trainingneg_dis.npy & MAE_dynamic_trainingpos_dis.npy

  • Initialization of the first unsupervised distance calculation in the dynamic dgmil algorithm.
  • Generated from the original features using the "get_score_and_dis_feats" function, as the first iteration requires initialization.

Overview of DGMIL

  • DGMIL is an instance-based Multiple Instance Learning (MIL) paradigm.
  • It trains at the instance level and during testing, obtains predictions for each instance, followed by aggregation methods like mean-pooling or max-pooling to derive bag predictions.

Additional Tips

  • The quality of the original feature space significantly impacts DGMIL, likely due to distance measurement and clustering algorithms.
  • Through extensive experimentation, it's found that MAE features might be more favorable for DGMIL compared to Simclr and ImageNet pretrained features.

Frequently Asked Questions.

  • Regarding the different result with DSMIL.

    Compared with DSMIL, for considerations of computational efficiency and resources, we used 5x (vs. DSMIL 20x) in our experiments. We used a patch size of 512 (vs DSMIL 224), and a patch is labeled as positive if it contains 25% or more cancer areas (not specified in DSMIL, please refer to its code). These different settings may result in the difference between the metrics reported by us and those reported by DSMIL.

  • Regarding the issues with the MAE model in the article.

    As mentioned in the article, we used the official MAE model to pre-train a feature extractor for pre-extracting features of all patches (since the number of patches cut from WSI is too large, direct end-to-end training based on the RGB image is too costly, and the existing common methods all adopt pre-trained feature extractors for feature extraction). Subsequently, we only trained a simple projector (inspired by SimCLR to complete the mapping of the feature space, while the pre-extracted features remain unchanged. This can also be understood as the overall feature extractor including a pre-trained but fixed MAE model during training, and a projector that is updated during training.

    For specific implementations, as the settings of different MIL experiments vary (such as patch size, scale, etc.), patching needs to be conducted according to your own experimental settings. The DSMIL paper provides a good example for reference (and is also referenced in this article). For MAE self-supervised feature extraction, you can refer to MAE's official code, and many pre-trained models are available for use. In this paper, an MAE model was trained on the dataset after patching using the official MAE code.

    As uploading all these extracted feats files would require a lot of time and space, we have open-sourced the main and key code models. The guidance in the readme and the above guidance can support the reproduction of this work. Thank you again for your attention! You are welcome to contact and cite us! Thank you!

  • Regarding the multi-class classification task.

    Currently, dealing with multi-class classification problems, you can transform them into multiple binary problems to solve.

Citation

If this work is helpful to you, please cite it as:

@InProceedings{10.1007/978-3-031-16434-7_3,
author="Qu, Linhao
and Luo, Xiaoyuan
and Liu, Shaolei
and Wang, Manning
and Song, Zhijian",
title="DGMIL: Distribution Guided Multiple Instance Learning for Whole Slide Image Classification",
booktitle="Medical Image Computing and Computer Assisted Intervention -- MICCAI 2022",
year="2022",
publisher="Springer Nature Switzerland",
address="Cham",
pages="24--34",
isbn="978-3-031-16434-7"
}

Contact Information

If you have any question, please email to me lhqu20@fudan.edu.cn.

About

Official PyTorch implementation of our MICCAI 2022 paper: DGMIL: Distribution Guided Multiple Instance Learning for Whole Slide Image Classification.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages