Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1]
Additive MIL: Intrinsically Interpretable Multiple Instance Learning for Pathology
Authors: Syed Ashar Javed, Dinkar Juyal, Harshith Padigela, Amaro Taylor-Weiner, Limin Yu, Aaditya Prakash
NeurIPS 2022 | Venue PDF | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | We perform various experiments to show the benefits of using Additive MIL models for interpretability in pathology problems. |
| Researcher Affiliation | Industry | Syed Ashar Javed Path AI Inc EMAIL Dinkar Juyal Path AI Inc EMAIL Harshith Padigela Path AI Inc EMAIL Amaro Taylor-Weiner Path AI Inc EMAIL Limin Yu Path AI Inc EMAIL Aaditya Prakash Path AI Inc EMAIL |
| Pseudocode | No | The paper does not contain any pseudocode or clearly labeled algorithm blocks. |
| Open Source Code | Yes | Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] |
| Open Datasets | Yes | The first problem is the prediction of cancer subtypes in non-small cell lung carcinoma (NSCLC) and renal cell carcinoma (RCC), both of which use the TCGA dataset [45]. The second problem is the detection of metastasis in breast cancer using the Camelyon16 dataset [6]. |
| Dataset Splits | Yes | Both TCGA datasets were split into 60/15/25 (train/val/test) as done previously [36] while ensuring no data leakage at a case level. |
| Hardware Specification | Yes | All training and inference runs were done on Quadro RTX 8000, and it takes 3 to 4 hours to train the model with four GPUs. |
| Software Dependencies | No | The paper mentions 'ADAM optimizer' and 'Shufflenet' but does not provide specific version numbers for these or any other software dependencies. |
| Experiment Setup | Yes | For training the models, a bag size of 48-1600 patches and batch size of 16-64 was experimented with and the best one chosen using cross-validation. [...] the entire model was trained with ADAM optimizer [21] and a learning rate of 1e-4. |