pith. machine review for the scientific record. sign in

arxiv: 2602.18502 · v2 · submitted 2026-02-17 · 💻 cs.CV · cs.LG

Recognition: no theorem link

Mitigating Shortcut Learning via Feature Disentanglement in Medical Imaging: A Benchmark Study

Authors on Pith no claims yet

Pith reviewed 2026-05-15 21:55 UTC · model grok-4.3

classification 💻 cs.CV cs.LG
keywords shortcut learningfeature disentanglementmedical imagingadversarial learninglatent space analysisconfounding factorsrobustnessbenchmark study
0
0 comments X

The pith

Combining feature disentanglement with data rebalancing mitigates shortcut learning more robustly than rebalancing alone in medical imaging.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper evaluates feature disentanglement methods, including adversarial learning and latent space splitting via dependence minimization, as a way to separate task-relevant features from confounding factors in deep learning models for medical image classification. These approaches target the problem of models exploiting spurious correlations that fail to generalize across hospitals, populations, or scanners. Tests on one artificial dataset and two medical datasets with both natural and synthetic confounders show that pairing disentanglement with data rebalancing improves performance specifically when spurious correlations are strong in training data. Latent space analyses expose representation differences that classification accuracy alone does not reveal, and model reliance on shortcuts scales with the degree of confounding present. The combined strategy delivers stronger shortcut mitigation than rebalancing by itself at comparable computational cost.

Core claim

The study establishes that the best-performing models integrate data-centric rebalancing with model-centric disentanglement to achieve stronger and more robust shortcut mitigation than rebalancing alone, while preserving similar computational efficiency. This outcome holds across datasets that vary in confounding strength, with latent space metrics showing that each disentanglement technique produces distinct representation qualities not captured by accuracy scores.

What carries the argument

Feature disentanglement through adversarial learning and dependence-minimizing latent space splitting, which isolates task-relevant information from confounder-related features in the model's latent representations.

If this is right

  • Classification performance improves under strong spurious correlations in the training data.
  • Latent space analyses reveal representation quality differences not visible from classification metrics alone.
  • Model reliance on shortcuts increases as the degree of confounding in training data rises.
  • The combined approach maintains computational efficiency comparable to rebalancing alone.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The methods could extend to other imaging domains such as radiology or digital pathology where similar acquisition confounders appear.
  • Larger multi-center clinical validation would test whether the isolated features remain causal outside the studied datasets.
  • Standardizing the latent space metrics used here could enable direct comparisons of shortcut mitigation techniques across future studies.

Load-bearing premise

The chosen disentanglement methods and latent space metrics reliably isolate causally relevant features from non-causal confounders.

What would settle it

A new clinical dataset where the combined rebalancing-plus-disentanglement models show no gain in out-of-distribution accuracy or robustness over rebalancing alone would falsify the central claim.

Figures

Figures reproduced from arXiv: 2602.18502 by Philipp Berens, Sarah M\"uller.

Figure 1
Figure 1. Figure 1: Overview of shortcut learning and mitigation via feature disentanglement. [PITH_FULL_IMAGE:figures/full_fig_p002_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: Overview of label distributions in Morpho-MNIST, CheXpert, and OCT. a shows example images sampled for each label combina￾tion, b shows contingency tables of the original training data, and c shows contingency tables of the subsampled training data actually used. In the final training data (c), strong correlations were in￾duced between the primary task and confounder for all datasets, while maintaining bal… view at source ↗
Figure 4
Figure 4. Figure 4: Qualitative scatter plots showing the two-dimensional subspace [PITH_FULL_IMAGE:figures/full_fig_p011_4.png] view at source ↗
Figure 5
Figure 5. Figure 5: Relative AUROC improvement over the Baseline on the inverted test distribution ( [PITH_FULL_IMAGE:figures/full_fig_p012_5.png] view at source ↗
Figure 6
Figure 6. Figure 6: Disentanglement performance (diagonal dominance) of different methods as a function of [PITH_FULL_IMAGE:figures/full_fig_p013_6.png] view at source ↗
read the original abstract

Although deep learning models in medical imaging often achieve excellent classification performance, they can rely on shortcut learning, exploiting spurious correlations or confounding factors that are not causally related to the target task. This poses risks in clinical settings, where models must generalize across institutions, populations, and acquisition conditions. Feature disentanglement is a promising approach to mitigate shortcut learning by separating task-relevant information from confounder-related features in latent representations. In this study, we systematically evaluated feature disentanglement methods for mitigating shortcuts in medical imaging, including adversarial learning and latent space splitting based on dependence minimization. We assessed classification performance and disentanglement quality using latent space analyses across one artificial and two medical datasets with natural and synthetic confounders. We also examined robustness under varying levels of confounding and compared computational efficiency across methods. We found that shortcut mitigation methods improved classification performance under strong spurious correlations during training. Latent space analyses revealed differences in representation quality not captured by classification metrics, highlighting the strengths and limitations of each method. Model reliance on shortcuts depended on the degree of confounding in the training data. The best-performing models combine data-centric rebalancing with model-centric disentanglement, achieving stronger and more robust shortcut mitigation than rebalancing alone while maintaining similar computational efficiency. The project code is publicly available at https://github.com/berenslab/medical-shortcut-mitigation.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 1 minor

Summary. The manuscript reports a benchmark study evaluating feature disentanglement methods (adversarial learning and latent-space splitting via dependence minimization) for mitigating shortcut learning in medical image classification. On one artificial and two medical datasets with natural and synthetic confounders, the authors measure classification accuracy, latent-space quality metrics, robustness across confounding strengths, and runtime. They conclude that hybrid data-rebalancing plus disentanglement outperforms rebalancing alone while preserving efficiency, with code released publicly.

Significance. If the empirical findings hold under rigorous validation, the work supplies a practical benchmark showing that hybrid data-centric and model-centric interventions can improve robustness to spurious correlations in medical imaging without added computational cost. Public code supports reproducibility. The contribution is tempered by the absence of direct causal-feature recovery tests, limiting claims that observed gains reflect true isolation of causally relevant features rather than dataset-specific correlations.

major comments (2)
  1. Abstract and Results sections: The central claim that combining rebalancing with disentanglement yields stronger, more robust shortcut mitigation depends on the adversarial and dependence-minimization methods actually isolating causally relevant features from confounders. The evaluation relies on proxy metrics (dependence scores, latent-space quality) without ground-truth causal structure or direct recovery validation, leaving open the possibility that robustness gains reflect dataset-specific correlations instead of generalizable disentanglement.
  2. Experimental Setup (implied in Abstract): No details are supplied on statistical testing procedures, exact train/validation/test splits, or controls for selection effects in the benchmark datasets. These omissions are load-bearing for interpreting whether reported performance differences under varying confounding levels are statistically reliable.
minor comments (1)
  1. The description of latent-space analyses would benefit from explicit equations for the dependence-minimization objective and clearer notation distinguishing the different disentanglement losses.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the detailed and constructive review of our manuscript on mitigating shortcut learning via feature disentanglement in medical imaging. We address each major comment below and describe the revisions we plan to implement.

read point-by-point responses
  1. Referee: [—] Abstract and Results sections: The central claim that combining rebalancing with disentanglement yields stronger, more robust shortcut mitigation depends on the adversarial and dependence-minimization methods actually isolating causally relevant features from confounders. The evaluation relies on proxy metrics (dependence scores, latent-space quality) without ground-truth causal structure or direct recovery validation, leaving open the possibility that robustness gains reflect dataset-specific correlations instead of generalizable disentanglement.

    Authors: We agree that proxy metrics do not constitute direct causal validation, which is a known challenge in real-world medical datasets lacking ground-truth causal graphs. Our benchmark demonstrates empirical improvements in robustness and performance under controlled confounding variations, supported by latent space analyses. To address this, we will expand the Discussion section to explicitly discuss the limitations of proxy-based evaluation and the possibility of dataset-specific effects, while highlighting that the hybrid method's advantages hold across multiple datasets and confounding strengths. No direct causal recovery experiments will be added as they fall outside the scope of this benchmark study. revision: partial

  2. Referee: [—] Experimental Setup (implied in Abstract): No details are supplied on statistical testing procedures, exact train/validation/test splits, or controls for selection effects in the benchmark datasets. These omissions are load-bearing for interpreting whether reported performance differences under varying confounding levels are statistically reliable.

    Authors: We regret that these details were not sufficiently prominent in the main text. The full experimental protocol, including 5-fold cross-validation with stratified splits (70% train, 15% validation, 15% test), multiple random seeds for reproducibility, and statistical significance testing via paired t-tests (p < 0.05) with Bonferroni correction, is documented in the supplementary material and the released code. We will add a new subsection 'Experimental Details' in the Methods section to include this information explicitly, along with descriptions of how selection effects were controlled through balanced sampling and repeated experiments. revision: yes

Circularity Check

0 steps flagged

No circularity: purely empirical benchmark with no derivation chain

full rationale

This paper reports experimental results from training and evaluating disentanglement methods on three datasets, measuring classification accuracy, latent-space metrics, and robustness to confounding levels. No equations, fitted parameters presented as predictions, uniqueness theorems, or self-citation chains appear in the provided text. All central claims rest on observed performance differences rather than reducing to inputs by construction.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 0 invented entities

Empirical benchmark relying on standard machine-learning assumptions about representation learning and the existence of measurable confounders; no free parameters or invented entities introduced.

axioms (1)
  • domain assumption Feature disentanglement methods can separate task-relevant information from confounder-related features in latent representations
    Core premise underlying all tested mitigation strategies.

pith-pipeline@v0.9.0 · 5536 in / 1128 out tokens · 38433 ms · 2026-05-15T21:55:26.163143+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

82 extracted references · 82 canonical work pages · 6 internal anchors

  1. [1]

    The subgroup imperative: chest radiograph classifier generalization gaps in patient, setting, and pathology subgroups.Radiology: Artificial Intelligence, 5(5):e220270, 2023

    Monish Ahluwalia, Mohamed Abdalla, James Sanayei, Laleh Seyyed-Kalantari, Mohannad Hussain, Amna Ali, and Benjamin Fine. The subgroup imperative: chest radiograph classifier generalization gaps in patient, setting, and pathology subgroups.Radiology: Artificial Intelligence, 5(5):e220270, 2023

  2. [2]

    Achieving multisite generalizationforCNN-baseddiseasediagnosismodelsbymitigatingshortcutlearning.IEEE Access, 10:78726–78738, 2022

    Kaoutar Ben Ahmed, Lawrence O Hall, Dmitry B Goldgof, and Ryan Fogarty. Achieving multisite generalizationforCNN-baseddiseasediagnosismodelsbymitigatingshortcutlearning.IEEE Access, 10:78726–78738, 2022

  3. [3]

    Finding and removing clever hans: Using explanation methods to debug and improve deep models.Information Fusion, 77:261–295, 2022

    Christopher J Anders, Leander Weber, David Neumann, Wojciech Samek, Klaus-Robert Müller, and Sebastian Lapuschkin. Finding and removing clever hans: Using explanation methods to debug and improve deep models.Information Fusion, 77:261–295, 2022

  4. [4]

    Invariant Risk Minimization

    Martin Arjovsky, Léon Bottou, Ishaan Gulrajani, and David Lopez-Paz. Invariant risk minimization. arXiv preprint arXiv:1907.02893, 2019

  5. [5]

    Masktune: Mitigating spurious correlations by forcing to explore.Advances in Neural Information Processing Systems, 35:23284–23296, 2022

    Saeid Asgari, Aliasghar Khani, Fereshte Khani, Ali Gholami, Linh Tran, Ali Mahdavi Amiri, and Ghassan Hamarneh. Masktune: Mitigating spurious correlations by forcing to explore.Advances in Neural Information Processing Systems, 35:23284–23296, 2022

  6. [6]

    Explanation is all you need in distillation: Mitigating bias and shortcut learning.arXiv preprint arXiv:2407.09788, 2024

    Pedro RAS Bassi, Andrea Cavalli, and Sergio Decherchi. Explanation is all you need in distillation: Mitigating bias and shortcut learning.arXiv preprint arXiv:2407.09788, 2024

  7. [7]

    Mutual Information Neural Estimation

    Mohamed Ishmael Belghazi, Aristide Baratin, Sai Rajeshwar, Sherjil Ozair, Yoshua Bengio, Aaron Courville, and Devon Hjelm. Mutual Information Neural Estimation. InProceedings of the 35th International Conference on Machine Learning, volume 80 ofProceedings of Machine Learning Research, pages 531–540. PMLR, 2018

  8. [8]

    All you need is a guiding hand: Mitigating shortcut bias in deep learning models for medical imaging

    Christopher Boland, Owen Anderson, Keith A Goatman, John Hipwell, Sotirios A Tsaftaris, and Sonia Dahdouh. All you need is a guiding hand: Mitigating shortcut bias in deep learning models for medical imaging. InMICCAI Workshop on Fairness of AI in Medical Imaging, pages 67–77. Springer, 2024

  9. [9]

    Preventing shortcut learning in medical image analysis through intermediate layer knowledge distillation from specialist teachers.arXiv preprint arXiv:2511.17421, 2025

    Christopher Boland, Sotirios Tsaftaris, and Sonia Dahdouh. Preventing shortcut learning in medical image analysis through intermediate layer knowledge distillation from specialist teachers.arXiv preprint arXiv:2511.17421, 2025

  10. [10]

    Detecting shortcut learning for fair medical ai using shortcut testing.Nature communica- tions, 14(1):4314, 2023

    Alexander Brown, Nenad Tomasev, Jan Freyberg, Yuan Liu, Alan Karthikesalingam, and Jessica Schrouff. Detecting shortcut learning for fair medical ai using shortcut testing.Nature communica- tions, 14(1):4314, 2023

  11. [11]

    Castro, Jeremy Tan, Bernhard Kainz, Ender Konukoglu, and Ben Glocker

    Daniel C. Castro, Jeremy Tan, Bernhard Kainz, Ender Konukoglu, and Ben Glocker. Morpho- mnist: Quantitative Assessment and Diagnostics for Representation Learning.Journal of Machine Learning Research, 20(178):1–29, 2019. URLhttp://jmlr.org/papers/v20/19-033.html

  12. [12]

    Castro, Ian Walker, and Ben Glocker

    Daniel C. Castro, Ian Walker, and Ben Glocker. Causality matters in medical imaging.Nature Communications, 11(1), 2020. doi: 10.1038/s41467-020-17478-w

  13. [13]

    Domain generalization by mutual- information regularization with pre-trained models

    Junbum Cha, Kyungjae Lee, Sungrae Park, and Sanghyuk Chun. Domain generalization by mutual- information regularization with pre-trained models. InEuropean conference on computer vision, pages 440–457. Springer, 2022

  14. [14]

    International retrospective obser- vational study of continual learning for ai on endotracheal tube placement from chest radiographs

    Emma Chen, Agustina Saenz, Oishi Banerjee, Henrik Marklund, Xiaoman Zhang, Shreya Johri, Hong-Yu Zhou, Luyang Luo, Subathra Adithan, Kay Wu, et al. International retrospective obser- vational study of continual learning for ai on endotracheal tube placement from chest radiographs. NEJM AI, 3(1):AIoa2500522, 2025

  15. [15]

    When does group invariant learning survive spurious correlations?Advances in Neural Information Processing Systems, 35:7038–7051, 2022

    Yimeng Chen, Ruibin Xiong, Zhi-Ming Ma, and Yanyan Lan. When does group invariant learning survive spurious correlations?Advances in Neural Information Processing Systems, 35:7038–7051, 2022. 15

  16. [16]

    Ai act | shaping europe’s digital future.https://digital-strategy.ec

    European Commission. Ai act | shaping europe’s digital future.https://digital-strategy.ec. europa.eu/en/policies/regulatory-framework-ai, 2024. Accessed January 2026

  17. [17]

    Ai for radiographic covid-19 detection selects shortcuts over signal.Nature Machine Intelligence, 3(7):610–619, 2021

    Alex J DeGrave, Joseph D Janizek, and Su-In Lee. Ai for radiographic covid-19 detection selects shortcuts over signal.Nature Machine Intelligence, 3(7):610–619, 2021

  18. [18]

    PyTorch Lightning, March 2019

    William Falcon and The PyTorch Lightning team. PyTorch Lightning, March 2019. URLhttps: //github.com/Lightning-AI/lightning

  19. [19]

    Avoiding Shortcut- Learning by Mutual Information Minimization in Deep Learning-Based Image Processing.IEEE Access, 11:64070–64086, 2023

    Louisa Fay, Erick Cobos, Bin Yang, Sergios Gatidis, and Thomas Küstner. Avoiding Shortcut- Learning by Mutual Information Minimization in Deep Learning-Based Image Processing.IEEE Access, 11:64070–64086, 2023. doi: 10.1109/ACCESS.2023.3289397

  20. [20]

    Mimm-x: Disentan- gling spurious correlations for medical image analysis

    Louisa Fay, Hajer Reguigui, Bin Yang, Sergios Gatidis, and Thomas Küstner. Mimm-x: Disentan- gling spurious correlations for medical image analysis. InMICCAI Workshop on Fairness of AI in Medical Imaging, pages 94–103. Springer, 2025

  21. [21]

    Food and Drug Administration

    U.S. Food and Drug Administration. Artificial intelligence in software as a medi- cal device.https://www.fda.gov/medical-devices/software-medical-device-samd/ artificial-intelligence-software-medical-device, 2026. Accessed January 2026

  22. [22]

    Unsupervised domain adaptation by backpropagation

    Yaroslav Ganin and Victor Lempitsky. Unsupervised domain adaptation by backpropagation. In Proceedings of the 32nd International Conference on International Conference on Machine Learning - Volume 37, pages 1180–1189, 2015

  23. [23]

    Shortcut learning in deep neural networks.Nature Machine Intelligence, 2(11):665–673, 2020

    Robert Geirhos, Jörn-Henrik Jacobsen, Claudio Michaelis, Richard Zemel, Wieland Brendel, Matthias Bethge, and Felix A Wichmann. Shortcut learning in deep neural networks.Nature Machine Intelligence, 2(11):665–673, 2020

  24. [24]

    Borgwardt, Malte J

    Arthur Gretton, Karsten M. Borgwardt, Malte J. Rasch, Bernhard Schölkopf, and Alexander Smola. A kernel two-sample test.Journal of Machine Learning Research, 13(25):723–773, 2012. URL http://jmlr.org/papers/v13/gretton12a.html

  25. [25]

    Adver- sarial domain adaptation network for tumor image diagnosis.International Journal of Approximate Reasoning, 135:38–52, 2021

    Chunmei He, Shunmin Wang, Hongyu Kang, Lanqing Zheng, Taifeng Tan, and Xianjun Fan. Adver- sarial domain adaptation network for tumor image diagnosis.International Journal of Approximate Reasoning, 135:38–52, 2021

  26. [26]

    Distilling the Knowledge in a Neural Network

    Geoffrey Hinton, Oriol Vinyals, and Jeff Dean. Distilling the knowledge in a neural network.arXiv preprint arXiv:1503.02531, 2015

  27. [27]

    Simple data balancing achieves competitive worst-group-accuracy

    Badr Youbi Idrissi, Martin Arjovsky, Mohammad Pezeshki, and David Lopez-Paz. Simple data balancing achieves competitive worst-group-accuracy. InConference on Causal Learning and Rea- soning, pages 336–351. PMLR, 2022

  28. [28]

    Mong, Safwan S

    Jeremy Irvin, Pranav Rajpurkar, Michael Ko, Yifan Yu, Silviana Ciurea-Ilcus, Chris Chute, Hen- rik Marklund, Behzad Haghgoo, Robyn Ball, Katie Shpanskaya, Jayne Seekins, David A. Mong, Safwan S. Halabi, Jesse K. Sandberg, Ricky Jones, David B. Larson, Curtis P. Langlotz, Bhavik N. Patel, Matthew P. Lungren, and Andrew Y. Ng. Chexpert: A Large Chest Radiog...

  29. [29]

    On feature learning in the presence of spurious correlations.Advances in Neural Information Processing Systems, 35:38516– 38532, 2022

    Pavel Izmailov, Polina Kirichenko, Nate Gruver, and Andrew G Wilson. On feature learning in the presence of spurious correlations.Advances in Neural Information Processing Systems, 35:38516– 38532, 2022

  30. [30]

    Unsupervised domain adaptation in brain lesion segmentation with adversarial networks

    Konstantinos Kamnitsas, Christian Baumgartner, Christian Ledig, Virginia Newcombe, Joanna Simpson, Andrew Kane, David Menon, Aditya Nori, Antonio Criminisi, Daniel Rueckert, et al. Unsupervised domain adaptation in brain lesion segmentation with adversarial networks. InInter- national conference on information processing in medical imaging, pages 597–609....

  31. [31]

    Adaptive group robust ensemble knowledge distillation.arXiv preprint arXiv:2411.14984, 2024

    Patrik Kenfack, Ulrich Aïvodji, and Samira Ebrahimi Kahou. Adaptive group robust ensemble knowledge distillation.arXiv preprint arXiv:2411.14984, 2024

  32. [32]

    Labeled Optical Coherence Tomography (OCT) and Chest X-Ray Images for Classification, 2018

    Daniel Kermany, Kang Zhang, and Michael Goldbaum. Labeled Optical Coherence Tomography (OCT) and Chest X-Ray Images for Classification, 2018. URLhttps://data.mendeley.com/ datasets/rscbjbr9sj/2. 16

  33. [33]

    Last layer re-training is sufficient for robustness to spurious correlations.arXiv preprint arXiv:2204.02937, 2022

    Polina Kirichenko, Pavel Izmailov, and Andrew Gordon Wilson. Last layer re-training is sufficient for robustness to spurious correlations.arXiv preprint arXiv:2204.02937, 2022

  34. [34]

    Retraining an open-source pneumothorax detecting ma- chine learning algorithm for improved performance to medical images.Clinical imaging, 61:15–19, 2020

    Gene Kitamura and Christopher Deible. Retraining an open-source pneumothorax detecting ma- chine learning algorithm for improved performance to medical images.Clinical imaging, 61:15–19, 2020

  35. [35]

    Distribution shift detection for the postmarket surveillance of medical ai algorithms: a retrospective simulation study.npj Digital Medicine, 2024

    Lisa M Koch, Christian F Baumgartner, and Philipp Berens. Distribution shift detection for the postmarket surveillance of medical ai algorithms: a retrospective simulation study.npj Digital Medicine, 2024. doi: 10.1038/s41746-024-01085-w

  36. [36]

    Concept bottleneck models

    Pang Wei Koh, Thao Nguyen, Yew Siang Tang, Stephen Mussmann, Emma Pierson, Been Kim, and Percy Liang. Concept bottleneck models. InInternational conference on machine learning, pages 5338–5348. PMLR, 2020

  37. [37]

    Wilds: A benchmark of in-the-wild distribution shifts

    Pang Wei Koh, Shiori Sagawa, Henrik Marklund, Sang Michael Xie, Marvin Zhang, Akshay Bal- subramani, Weihua Hu, Michihiro Yasunaga, Richard Lanas Phillips, Irena Gao, et al. Wilds: A benchmark of in-the-wild distribution shifts. InInternational conference on machine learning, pages 5637–5664. PMLR, 2021

  38. [38]

    Learning to detour: Short- cut mitigating augmentation for weakly supervised semantic segmentation

    JuneHyoung Kwon, Eunju Lee, Yunsung Cho, and YoungBin Kim. Learning to detour: Short- cut mitigating augmentation for weakly supervised semantic segmentation. InProceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pages 819–828, 2024

  39. [39]

    Learning debiased representation via disentangled feature augmentation.Advances in Neural Information Processing Systems, 34:25123–25133, 2021

    Jungsoo Lee, Eungyeup Kim, Juyoung Lee, Jihyeon Lee, and Jaegul Choo. Learning debiased representation via disentangled feature augmentation.Advances in Neural Information Processing Systems, 34:25123–25133, 2021

  40. [40]

    Repair: Removing representation bias by dataset resampling

    Yi Li and Nuno Vasconcelos. Repair: Removing representation bias by dataset resampling. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 9572– 9581, 2019

  41. [41]

    Just train twice: Improving group robustness without training group information

    Evan Z Liu, Behzad Haghgoo, Annie S Chen, Aditi Raghunathan, Pang Wei Koh, Shiori Sagawa, Percy Liang, and Chelsea Finn. Just train twice: Improving group robustness without training group information. InInternational Conference on Machine Learning, pages 6781–6792. PMLR, 2021

  42. [42]

    The Variational Fair Autoencoder

    Christos Louizos, Kevin Swersky, Yujia Li, Max Welling, and Richard Zemel. The variational fair autoencoder.arXiv preprint arXiv:1511.00830, 2015

  43. [43]

    Pseudo bias-balanced learning for debiased chest x-ray classification

    Luyang Luo, Dunyuan Xu, Hao Chen, Tien-Tsin Wong, and Pheng-Ann Heng. Pseudo bias-balanced learning for debiased chest x-ray classification. InInternational conference on medical image com- puting and computer-assisted intervention, pages 621–631. Springer, 2022

  44. [44]

    Gen- erative interventions for causal learning

    Chengzhi Mao, Augustine Cha, Amogh Gupta, Hao Wang, Junfeng Yang, and Carl Vondrick. Gen- erative interventions for causal learning. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 3947–3956, 2021

  45. [45]

    Koch, Sergios Gatidis, Thomas Küstner, and Philipp Berens

    Sarah Müller, Louisa Fay, Lisa M. Koch, Sergios Gatidis, Thomas Küstner, and Philipp Berens. Benchmarking Dependence Measures to Prevent Shortcut Learning in Medical Imaging. InMachine Learning in Medical Imaging, pages 53–62. Springer Nature Switzerland, 2025. ISBN 978-3-031- 73290-4

  46. [46]

    Disentangling representations of retinal images with generative models.Medical Image Analysis, 105:103628, 2025

    Sarah Müller, Lisa M Koch, Hendrik PA Lensch, and Philipp Berens. Disentangling representations of retinal images with generative models.Medical Image Analysis, 105:103628, 2025

  47. [47]

    Uncovering and correcting shortcut learning in machine learning models for skin cancer diagnosis.Diagnostics, 12(1):40, 2021

    Meike Nauta, Ricky Walsh, Adam Dubowski, and Christin Seifert. Uncovering and correcting shortcut learning in machine learning models for skin cancer diagnosis.Diagnostics, 12(1):40, 2021

  48. [48]

    Simple disentan- glement of style and content in visual representations

    Lilian Ngweta, Subha Maity, Alex Gittens, Yuekai Sun, and Mikhail Yurochkin. Simple disentan- glement of style and content in visual representations. InInternational Conference on Machine Learning, pages 26063–26086. PMLR, 2023. 17

  49. [49]

    Decompose-and-compose: A compositional approach to miti- gating spurious correlation

    Fahimeh Hosseini Noohdani, Parsa Hosseini, Aryan Yazdan Parast, Hamidreza Yaghoubi Araghi, and Mahdieh Soleymani Baghshah. Decompose-and-compose: A compositional approach to miti- gating spurious correlation. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 27662–27671, 2024

  50. [50]

    Ethicsandgovernanceofartificialintelligenceforhealth: Whoguidance

    WorldHealthOrganization. Ethicsandgovernanceofartificialintelligenceforhealth: Whoguidance. https://www.who.int/publications/i/item/9789240037403, 2025. Accessed January 2026

  51. [51]

    Finding and fixing spurious patterns with explanations.arXiv preprint arXiv:2106.02112, 2021

    Gregory Plumb, Marco Tulio Ribeiro, and Ameet Talwalkar. Finding and fixing spurious patterns with explanations.arXiv preprint arXiv:2106.02112, 2021

  52. [52]

    New patch-based strategy for COVID-19 automatic identification using chest x-ray images.Health and Technology, 12(6):1117–1132, 2022

    Jorge A Portal-Diaz, Orlando Lovelle-Enríquez, Marlen Perez-Diaz, José D Lopez-Cabrera, Osmany Reyes-Cardoso, and Ruben Orozco-Morales. New patch-based strategy for COVID-19 automatic identification using chest x-ray images.Health and Technology, 12(6):1117–1132, 2022

  53. [53]

    Right for the Right Reasons: Training Differentiable Models by Constraining their Explanations

    Andrew Slavin Ross, Michael C Hughes, and Finale Doshi-Velez. Right for the right reasons: Training differentiable models by constraining their explanations.arXiv preprint arXiv:1703.03717, 2017

  54. [54]

    Distributionally Robust Neural Networks for Group Shifts: On the Importance of Regularization for Worst-Case Generalization

    Shiori Sagawa, Pang Wei Koh, Tatsunori B Hashimoto, and Percy Liang. Distributionally robust neural networks for group shifts: On the importance of regularization for worst-case generalization. arXiv preprint arXiv:1911.08731, 2019

  55. [55]

    An investigation of why overparameterization exacerbates spurious correlations

    Shiori Sagawa, Aditi Raghunathan, Pang Wei Koh, and Percy Liang. An investigation of why overparameterization exacerbates spurious correlations. InInternational Conference on Machine Learning, pages 8346–8356. PMLR, 2020

  56. [56]

    Making deep neural net- works right for the right scientific reasons by interacting with their explanations.Nature Machine Intelligence, 2(8):476–486, 2020

    Patrick Schramowski, Wolfgang Stammer, Stefano Teso, Anna Brugger, Franziska Herbert, Xiaoting Shao, Hans-Georg Luigs, Anne-Katrin Mahlein, and Kristian Kersting. Making deep neural net- works right for the right scientific reasons by interacting with their explanations.Nature Machine Intelligence, 2(8):476–486, 2020

  57. [57]

    Information-theoretic bias reduction via causal view of spurious correlation

    Seonguk Seo, Joon-Young Lee, and Bohyung Han. Information-theoretic bias reduction via causal view of spurious correlation. InProceedings of the AAAI Conference on Artificial Intelligence, volume 36, pages 2180–2188, 2022

  58. [58]

    Right for better reasons: Training differentiable models by constraining their influence functions

    Xiaoting Shao, Arseny Skryagin, Wolfgang Stammer, Patrick Schramowski, and Kristian Kersting. Right for better reasons: Training differentiable models by constraining their influence functions. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pages 9533–9540, 2021

  59. [59]

    Generalizability challenges of mortality risk prediction models: A retrospective analysis on a multi-center database.PLOS digital health, 1 (4):e0000023, 2022

    Harvineet Singh, Vishwali Mhasawade, and Rumi Chunara. Generalizability challenges of mortality risk prediction models: A retrospective analysis on a multi-center database.PLOS digital health, 1 (4):e0000023, 2022

  60. [60]

    Neural concept binder

    Wolfgang Stammer, Antonia Wüst, David Steinmann, and Kristian Kersting. Neural concept binder. Advances in Neural Information Processing Systems, 37:71792–71830, 2024

  61. [61]

    Navigating shortcuts, spurious correlations, and confounders: From origins via detection to mitigation.arXiv preprint arXiv:2412.05152, 2024

    David Steinmann, Felix Divo, Maurice Kraus, Antonia Wüst, Lukas Struppek, Felix Friedrich, and Kristian Kersting. Navigating shortcuts, spurious correlations, and confounders: From origins via detection to mitigation.arXiv preprint arXiv:2412.05152, 2024

  62. [62]

    Lever- aging diffusion-based image variations for robust training on poisoned data

    LukasStruppek, MartinHentschel, CliftonPoth, DominikHintersdorf, andKristianKersting. Lever- aging diffusion-based image variations for robust training on poisoned data. InNeurIPS 2023 Work- shop on Backdoors in Deep Learning-The Good, the Bad, and the Ugly

  63. [63]

    Székely, Maria L

    Gábor J. Székely, Maria L. Rizzo, and Nail K. Bakirov. Measuring and testing dependence by correlation of distances.The Annals of Statistics, 35(6), 2007. doi: 10.1214/009053607000000505

  64. [64]

    Explanatory interactive machine learning

    Stefano Teso and Kristian Kersting. Explanatory interactive machine learning. InProceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society, pages 239–245, 2019

  65. [65]

    Distilling fair representations from fair teachers.IEEE Transactions on Big Data, 2024

    Huan Tian, Bo Liu, Tianqing Zhu, Wanlei Zhou, and Philip S Yu. Distilling fair representations from fair teachers.IEEE Transactions on Big Data, 2024. 18

  66. [66]

    Inconsistent performance of deep learning models on mammogram classification.Journal of the American College of Radiology, 17(6):796–803, 2020

    Xiaoqin Wang, Gongbo Liang, Yu Zhang, Hunter Blanton, Zachary Bessinger, and Nathan Jacobs. Inconsistent performance of deep learning models on mammogram classification.Journal of the American College of Radiology, 17(6):796–803, 2020

  67. [67]

    Navigate beyond shortcuts: Debiased learning through the lens of neural collapse

    Yining Wang, Junjie Sun, Chenyue Wang, Mi Zhang, and Min Yang. Navigate beyond shortcuts: Debiased learning through the lens of neural collapse. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 12322–12331, 2024

  68. [68]

    On the effect of key factors in spurious correlation: A theoretical perspective

    Yipei Wang and Xiaoqian Wang. On the effect of key factors in spurious correlation: A theoretical perspective. InInternational Conference on Artificial Intelligence and Statistics, pages 3745–3753. PMLR, 2024

  69. [69]

    Discover and cure: Concept-aware mitigation of spurious correlation

    Shirley Wu, Mert Yuksekgonul, Linjun Zhang, and James Zou. Discover and cure: Concept-aware mitigation of spurious correlation. InInternational Conference on Machine Learning, pages 37765– 37786. PMLR, 2023

  70. [70]

    FairPrune: Achieving Fairness Through Pruning for Dermatological Disease Diagnosis

    Yawen Wu, Dewen Zeng, Xiaowei Xu, Yiyu Shi, and Jingtong Hu. FairPrune: Achieving Fairness Through Pruning for Dermatological Disease Diagnosis. InMedical Image Computing and Computer Assisted Intervention – MICCAI 2022, pages 743–753. Springer Nature Switzerland, 2022. ISBN 978-3-031-16431-6

  71. [71]

    Mi 2 gan: Gen- erative adversarial network for medical image domain adaptation using mutual information con- straint

    Xinpeng Xie, Jiawei Chen, Yuexiang Li, Linlin Shen, Kai Ma, and Yefeng Zheng. Mi 2 gan: Gen- erative adversarial network for medical image domain adaptation using mutual information con- straint. InInternational Conference on Medical Image Computing and Computer-Assisted Interven- tion, pages 516–525. Springer, 2020

  72. [72]

    Tsaftaris

    Yuyang Xue, Junyu Yan, Raman Dutt, Fasih Haider, Jingshuai Liu, Steven McDonagh, and Sotirios A. Tsaftaris. BMFT: Achieving Fairness via Bias-Based Weight Masking Fine-Tuning. InEthics and Fairness in Medical Imaging, pages 98–108. Springer Nature Switzerland, 2025. ISBN 978-3-031-72787-0

  73. [73]

    Machine learning generalizability across healthcare settings: insights from multi-site covid-19 screening.NPJ digital medicine, 5(1):69, 2022

    Jenny Yang, Andrew AS Soltan, and David A Clifton. Machine learning generalizability across healthcare settings: insights from multi-site covid-19 screening.NPJ digital medicine, 5(1):69, 2022

  74. [74]

    Chroma-vae: Mit- igating shortcut learning with generative classifiers.Advances in Neural Information Processing Systems, 35:20351–20365, 2022

    Wanqian Yang, Polina Kirichenko, Micah Goldblum, and Andrew G Wilson. Chroma-vae: Mit- igating shortcut learning with generative classifiers.Advances in Neural Information Processing Systems, 35:20351–20365, 2022

  75. [75]

    Identifying spurious biases early in training through the lens of simplicity bias

    Yu Yang, Eric Gan, Gintare Karolina Dziugaite, and Baharan Mirzasoleiman. Identifying spurious biases early in training through the lens of simplicity bias. InInternational conference on artificial intelligence and statistics, pages 2953–2961. PMLR, 2024

  76. [76]

    Improving out-of-distribution robustness via selective augmentation

    HuaxiuYao, YuWang, SaiLi, LinjunZhang, WeixinLiang, JamesZou, andChelseaFinn. Improving out-of-distribution robustness via selective augmentation. InInternational Conference on Machine Learning, pages 25407–25437. PMLR, 2022

  77. [77]

    Cutmix: Regularization strategy to train strong classifiers with localizable features

    Sangdoo Yun, Dongyoon Han, Seong Joon Oh, Sanghyuk Chun, Junsuk Choe, and Youngjoon Yoo. Cutmix: Regularization strategy to train strong classifiers with localizable features. InProceedings of the IEEE/CVF international conference on computer vision, pages 6023–6032, 2019

  78. [78]

    Removal of confounders via invariant risk minimization for medical diagnosis

    Samira Zare and Hien Van Nguyen. Removal of confounders via invariant risk minimization for medical diagnosis. InInternational Conference on Medical Image Computing and Computer-Assisted Intervention, pages 578–587. Springer, 2022

  79. [79]

    mixup: Beyond Empirical Risk Minimization

    Hongyi Zhang, Moustapha Cisse, Yann N Dauphin, and David Lopez-Paz. mixup: Beyond empirical risk minimization.arXiv preprint arXiv:1710.09412, 2017

  80. [80]

    Adversarial hospital-invariant feature learning for WSI patch classification

    Mengliang Zhang. Adversarial hospital-invariant feature learning for WSI patch classification. In Submitted to Medical Imaging with Deep Learning, 2025. URLhttps://openreview.net/forum? id=R8k4P4IV14. under review

Showing first 80 references.