pith. machine review for the scientific record. sign in

arxiv: 2604.08936 · v1 · submitted 2026-04-10 · 💻 cs.CV

Recognition: unknown

M-IDoL: Information Decomposition for Modality-Specific and Diverse Representation Learning in Medical Foundation Model

Authors on Pith no claims yet

Pith reviewed 2026-05-10 17:14 UTC · model grok-4.3

classification 💻 cs.CV
keywords medical foundation modelinformation decompositionmodality-specific representationsmixture of expertsself-supervised learningmultimodal medical imagesclinical task generalization
0
0 comments X

The pith

A medical foundation model separates image modalities into distinct subspaces to reduce blending and improve task performance.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper presents M-IDoL as a self-supervised approach that decomposes information from multimodal medical images to avoid the common problem of blended representations. It achieves this through two objectives that push representations into separate Mixture-of-Experts subspaces for modality specificity while sharpening semantic differences inside each subspace for greater diversity. Pre-training occurs on 1.15 million images, after which the model shows stronger results than prior approaches on 21 clinical tasks spanning five modalities including X-ray, fundus, OCT, dermoscopy, and pathology. A reader would care because such separation could yield medical AI that handles varied scan types with less confusion and more reliable feature use.

Core claim

M-IDoL learns universal representations from multimodal medical images by maximizing inter-modality entropy to disperse them into separable MoE subspaces for specificity across modalities and minimizing intra-modality uncertainty via fine-grained semantic discrimination within each subspace to enrich diversity per modality. This produces clearer separation of feature clusters across modalities and finer discrimination within each, leading to superior generalization on 21 downstream clinical tasks and outperformance of 20 other foundation models on five imaging modalities.

What carries the argument

Mixture-of-Experts subspaces that enforce information decomposition by dispersing representations for inter-modality specificity and refining them for intra-modality diversity.

Load-bearing premise

The two information decomposition objectives will produce representations that generalize better without causing information loss or optimization instabilities.

What would settle it

A controlled experiment training the same model without the entropy maximization and uncertainty minimization objectives and measuring no gain or a drop in performance across the 21 clinical tasks would indicate the decomposition is not the source of improvement.

Figures

Figures reproduced from arXiv: 2604.08936 by Heng Tao Shen, Jiaxiong Yang, Lianghua He, Longzhen Yang, Yihang Liu, Ying Wen.

Figure 1
Figure 1. Figure 1: (a) Medical images exhibit inter-modality specificity and intra-modality diversity. (b) Modality-specific models excels in intra-modality diversity by focusing on stable imaging statistics. (c) MFMs suffer from information ambiguity due to uniform max￾imization of redundancy (Rdncy.) information across modalities. (d) M-IDoL mitigates ambiguity via information decomposition, enhancing modality-specific and… view at source ↗
Figure 2
Figure 2. Figure 2: Left: Overview of M-IDoL. Via information decomposition, M-IDoL optimizes two objectives: (a) the routing-consistency loss Lroute, which learns modality-separable MoE subspaces to maximize H(X|Z) for inter-modality specificity, and (b) the intra-modality contrastive loss Lcst, which promotes fine-grained discrimination within each modality to minimize H(X|Y, Z) for intra-modality diversity. Right: (a) Inte… view at source ↗
Figure 3
Figure 3. Figure 3: (a)Visualization of routing assignments for 1,000 images per modality. (b) Impact of expert number on downstream tasks. subspace is proportional to maximizing the expected log ratio: Ep(x,y,z)  log p(x|y) p(x)  → max, (18) which encourages the augmented representation Y to con￾tribute more semantics-invariant information that reduces the predictive uncertainty of X beyond the representation Z from other … view at source ↗
Figure 4
Figure 4. Figure 4: Comparison with unified MFMs. Error bars denote standard deviation,the center position reflects the mean performance score (↑, %). p < 0.05 indicates that our M-IDoL significantly outperforms the second-best method. Statistic results are provided in Appendix D. (a)RETFound (Fundus-specific) (b)AFiRE (X-ray-specific) (c)RETFound (OCT-specific) (f)Joint Pre-training (g) Medcoss (Unified) (h)CoSMIC (Unified) … view at source ↗
Figure 5
Figure 5. Figure 5: t-SNE clusters of representations across five modalities. (a–e) modality-specific models, (f–h) unified MFMs and (i) M-IDoL. on most datasets across modalities, and attains competitive performance on NIH (P = 0.099) and ISIC2016 (P = 0.328) compared to CoSMIC. Particularly, M-IDoL outperforms UniMiss+ (Xie et al., 2024), LVM-Med (MH Nguyen et al., 2024), and UniMed (Khattak et al., 2024) by more than 6%, h… view at source ↗
Figure 6
Figure 6. Figure 6: Loss Curves from Training Log. we apply random image augmentation techniques as same as DINO (Caron et al., 2021) such as Horizontal Flip, Color Jitter, Gaussian Blur, and Solarization. Architecture. The M-IDoL pre-training framework consists of two Swin Transformer-Base (Swin-B) visual encoders, namely a student encoder Sθ and a teacher encoder Tθ. Both encoders share the same Swin-B architecture: the inp… view at source ↗
Figure 7
Figure 7. Figure 7: Confusion-matrix of our proposed M-IDoL. 18 [PITH_FULL_IMAGE:figures/full_fig_p018_7.png] view at source ↗
Figure 8
Figure 8. Figure 8 [PITH_FULL_IMAGE:figures/full_fig_p019_8.png] view at source ↗
read the original abstract

Medical foundation models (MFMs) aim to learn universal representations from multimodal medical images that can generalize effectively to diverse downstream clinical tasks. However, most existing MFMs suffer from information ambiguity that blend multimodal representations in a single embedding space, leading to the degradation of modality specificity and diversity. In this paper, we propose M-IDoL, a self-supervised \underline{\textit{M}}FM that introduces Information Decomposition for multimodal representation Learning via two objectives: i) maximize inter-modality entropy by dispersing multimodal representation into separable Mixture-of-Experts (MoE) subspaces to achieve representation specificity across modalities; and ii) minimize intra-modality uncertainty by performing fine-grained semantic discrimination within each MoE subspace to enrich representation diversity per modality. By pre-training on 1.15 million medical images, M-IDoL i) delivers superior generalization across 21 downstream clinical tasks, outperforming 20 foundation models on five imaging modalities (e.g., X-ray, fundus, OCT, dermoscopy and pathology), and ii) learns modality-specific and diverse representations, showing clearer separation of feature cluster across modalities and finer-grained feature discrimination within each modality.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

0 major / 3 minor

Summary. The manuscript proposes M-IDoL, a self-supervised medical foundation model that performs information decomposition on multimodal medical images via two objectives: (i) maximizing inter-modality entropy by routing representations into separable Mixture-of-Experts (MoE) subspaces to enforce modality specificity, and (ii) minimizing intra-modality uncertainty through fine-grained semantic discrimination within each subspace to promote diversity. Pre-trained on 1.15 million images, the model claims superior generalization across 21 downstream clinical tasks, outperforming 20 existing foundation models on five modalities (X-ray, fundus, OCT, dermoscopy, pathology), while visualizations indicate clearer inter-modality cluster separation and finer intra-modality feature discrimination.

Significance. If the empirical gains are reproducible, the work meaningfully advances multimodal medical foundation models by explicitly addressing representation blending through MoE-based decomposition rather than relying on implicit separation in a shared embedding space. The scale of pre-training data and breadth of evaluation (21 tasks, 5 modalities) strengthen the case for practical utility. The internal consistency of the loss formulations and MoE routing mechanism, as confirmed by review of the full methods, supports the approach without introducing obvious optimization instabilities or information collapse.

minor comments (3)
  1. [Abstract] Abstract: The summary asserts 'superior generalization' and 'outperforming 20 foundation models' but omits any quantitative metrics, effect sizes, or statistical tests; while these appear in the results section, including one or two headline numbers (e.g., average improvement on the 21 tasks) would make the central claim immediately verifiable.
  2. [Section 3] Section 3 (Methods): The MoE routing and entropy objectives are described clearly, but the exact formulation of the semantic discrimination loss (e.g., whether it is a standard contrastive or classification loss) would benefit from an explicit equation reference to allow direct comparison with prior work.
  3. [Figure 4] Figure 4 (visualizations): The t-SNE or UMAP plots demonstrate the claimed separation, but the caption should specify the exact number of samples per modality and the perplexity or other hyperparameters used to ensure reproducibility of the qualitative evidence.

Simulated Author's Rebuttal

0 responses · 0 unresolved

We thank the referee for the positive and constructive review of our manuscript. We appreciate the recognition of the significance of M-IDoL in addressing representation blending in multimodal medical foundation models through explicit MoE-based information decomposition, as well as the acknowledgment of the scale of our pre-training and evaluation. The recommendation for minor revision is noted, and we will incorporate appropriate clarifications and improvements in the revised version.

Circularity Check

0 steps flagged

No significant circularity

full rationale

The paper presents an empirical self-supervised pre-training method using two information-decomposition objectives (inter-modality entropy maximization via MoE subspaces and intra-modality uncertainty minimization via semantic discrimination). All central claims of superior generalization on 21 downstream tasks and improved representation separation are validated through external benchmarks against 20 other models, not through any closed-form derivation or fitted parameter that reduces to the training inputs by construction. No equations are shown that equate predictions to inputs, no self-citation chains support uniqueness theorems, and the loss formulations are explicitly designed to target the stated goals without smuggling in the target performance metrics. The method is self-contained against the reported empirical outcomes.

Axiom & Free-Parameter Ledger

0 free parameters · 2 axioms · 1 invented entities

The central claim depends on the unverified effectiveness of the two novel decomposition objectives and the assumption that MoE subspaces can cleanly separate modalities without downstream cost; no free parameters or external benchmarks are described.

axioms (2)
  • domain assumption Self-supervised objectives on unlabeled medical images can produce transferable representations for downstream clinical tasks
    Standard premise in foundation model literature invoked by the pre-training description
  • ad hoc to paper Mixture-of-Experts subspaces can be trained to disperse multimodal representations without collapsing useful information
    Core mechanism introduced to achieve modality specificity
invented entities (1)
  • Modality-specific MoE subspaces for information decomposition no independent evidence
    purpose: To achieve representation specificity across modalities and diversity within each modality
    New construct proposed in the method; no independent evidence outside the paper's claims

pith-pipeline@v0.9.0 · 5517 in / 1417 out tokens · 51072 ms · 2026-05-10T17:14:12.254969+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

26 extracted references · 11 canonical work pages · 1 internal anchor

  1. [1]

    https://www.kaggle.com/datasets/ orvile/neh-ut-oct-dataset

    Neh-oct. https://www.kaggle.com/datasets/ orvile/neh-ut-oct-dataset. Kaggle. Retina dataset. https://www.kaggle.com/ datasets/jr2ngb/cataractdataset/data. Kaggle. Siim-acr pneumothorax segmentation. https://www.kaggle.com/c/ siim-acr-pneumothorax-segmentation . Kaggle. Bell, A. J. The co-information lattice. InProceedings of the Fifth International Worksh...

  2. [2]

    H., Pinto, U

    Cano, J. H., Pinto, U. O., and Th´ebault, S. Dataset of eye fundus and oct images for the study of diabetic macu- lar edema and diabetic retinopathy.Translational Visual Health Laboratory, Instituto de Neurobiolog´ıa, Universi- dad Nacional Aut´onoma de M´exico (UNAM), Quer´etaro, Mexico, Tech. Report CF-2019-1759 and IN, 205420,

  3. [3]

    arXiv preprint arXiv:2506.08356 (2025)

    Chopra, S., Sanchez-Rodriguez, G., Mao, L., Feola, A. J., Li, J., and Kira, Z. Medmoe: modality-specialized mixture of experts for medical vision-language understanding.arXiv preprint arXiv:2506.08356,

  4. [4]

    Skin Lesion Analysis Toward Melanoma Detection 2018: A Challenge Hosted by the International Skin Imaging Collaboration (ISIC)

    Codella, N., Rotemberg, V ., Tschandl, P., et al. Skin lesion analysis toward melanoma detection 2018: A challenge hosted by the international skin imaging collaboration (isic).arXiv preprint arXiv:1902.03368,

  5. [5]

    C., Gutman, D., Celebi, M

    Codella, N. C., Gutman, D., Celebi, M. E., et al. Skin lesion analysis toward melanoma detection: A challenge at the 2017 international symposium on biomedical imaging (isbi), hosted by the international skin imaging collabora- tion (isic). In2018 IEEE 15th international symposium on biomedical imaging (ISBI 2018), pp. 168–172,

  6. [6]

    Scaling self- supervised learning for histopathology with masked im- age modeling.MedRxiv, pp

    Filiot, A., Ghermi, R., Olivier, A., et al. Scaling self- supervised learning for histopathology with masked im- age modeling.MedRxiv, pp. 2023–07,

  7. [7]

    C., Celebi, E., et al

    Gutman, D., Codella, N. C., Celebi, E., et al. Skin lesion analysis toward melanoma detection: A challenge at the international symposium on biomedical imaging (isbi) 2016, hosted by the international skin imaging collabora- tion (isic).arXiv preprint arXiv:1605.01397,

  8. [8]

    Learning deep representations by mutual information estimation and maximization

    Hjelm, R. D., Fedorov, A., Lavoie-Marchildon, S., Grewal, K., Bachman, P., Trischler, A., and Bengio, Y . Learning deep representations by mutual information estimation and maximization.arXiv preprint arXiv:1808.06670,

  9. [9]

    arXiv:1901.07042 (2019) 10 A.Rafferty et al

    Irvin, J., Rajpurkar, P., Ko, M., et al. Chexpert: A large chest radiograph dataset with uncertainty labels and ex- pert comparison. InProceedings of the AAAI conference on artificial intelligence, volume 33(01), pp. 590–597, 2019a. Irvin, J., Rajpurkar, P., Ko, M., et al. Chexpert: A large chest radiograph dataset with uncertainty labels and ex- pert com...

  10. [10]

    Aptos 2019 blindness de- tection

    Karthik, Maggie, and Dane, S. Aptos 2019 blindness de- tection. https://kaggle.com/competitions/ aptos2019-blindness-detection,

  11. [11]

    arXiv preprint arXiv:2412.10372 (2024)

    Khattak, M. U., Kunhimon, S., Naseer, M., Khan, S., and Khan, F. S. Unimed-clip: Towards a unified image-text pretraining paradigm for diverse medical imaging modal- ities.arXiv preprint arXiv:2412.10372,

  12. [12]

    Kovalyk, O., Morales-S ´anchez, J., Verd ´u-Monedero, R., et al

    URL https: //doi.org/10.7910/DVN/1YRRAC. Kovalyk, O., Morales-S ´anchez, J., Verd ´u-Monedero, R., et al. Papila. https://doi.org/10.6084/m9. figshare.14798004.v1,

  13. [13]

    GShard: Scaling Giant Models with Conditional Computation and Automatic Sharding

    Lepikhin, D., Lee, H., Xu, Y ., et al. Gshard: Scaling gi- ant models with conditional computation and automatic sharding.arXiv preprint arXiv:2006.16668,

  14. [14]

    Noise Contrastive Estimation and Negative Sampling for Conditional Models: Consistency and Statistical Efficiency

    Ma, Z. and Collins, M. Noise contrastive estimation and neg- ative sampling for conditional models: Consistency and statistical efficiency.arXiv preprint arXiv:1809.01812, 2018a. Ma, Z. and Collins, M. Noise contrastive estimation and neg- ative sampling for conditional models: Consistency and statistical efficiency.arXiv preprint arXiv:1809.01812, 2018b....

  15. [15]

    K., Aktas, H

    Peltekian, A. K., Aktas, H. E., Durak, G., Grudzinski, K., Bemiss, B. C., Richardson, C., Dematte, J. E., Budinger, G., Esposito, A. J., Misharin, A., et al. Ren: Anatomically-informed mixture-of-experts for interstitial lung disease diagnosis.arXiv preprint arXiv:2510.04923,

  16. [16]

    K., GELLY, S., LUCIC, M

    Tschannen, M., Djolonga, J., Rubenstein, P. K., Gelly, S., and Lucic, M. On mutual information maximization for representation learning.arXiv preprint arXiv:1907.13625,

  17. [17]

    Chestx-ray8: Hospital- scale chest x-ray database and benchmarks on weakly- supervised classification and localization of common tho- rax diseases

    Wang, X., Peng, Y ., Lu, L., et al. Chestx-ray8: Hospital- scale chest x-ray database and benchmarks on weakly- supervised classification and localization of common tho- rax diseases. InProceedings of the IEEE conference on computer vision and pattern recognition, pp. 2097–2106,

  18. [18]

    Images are categorized as glaucomatous, non-glaucomatous, or suspect, based on comprehensive clinical evaluation

    contains 488 retinal images collected at HGURS (Murcia, Spain) between 2018–2020. Images are categorized as glaucomatous, non-glaucomatous, or suspect, based on comprehensive clinical evaluation. Retina(Ret) is established by Seoul National University (South Korea) to support automated retinal disease detection. It includes 601 images spanning four catego...

  19. [19]

    consists of 1,113 macular OCT images acquired between 2015 and 2022, intended for diagnosing DME or DR

  20. [20]

    Grand Challenge

    is a binary classification benchmark for colorectal polyp histology, comprising 3,152 hematoxylin-and-eosin (H&E) stained formalin-fixed, paraffin-embedded (FFPE) image patches of fixed size (224×224 pixels). Images are labeled as Hyperplastic Polyp (HP) or Sessile Serrated Adenoma (SSA). To address class imbalance, we applied random cropping and horizon-...

  21. [21]

    We use all images for training, and evaluate on the 1,512 image test set provided by the ISIC 2018 challenge

    consists of 10,015 der- matoscopic images across seven disease categories. We use all images for training, and evaluate on the 1,512 image test set provided by the ISIC 2018 challenge. ISIC2018(Codella et al.,

  22. [22]

    Architecture.The M-IDoL pre-training framework consists of two Swin Transformer-Base (Swin-B) visual encoders, namely a student encoder Sθ and a teacher encoder Tθ

    such as Horizontal Flip, Color Jitter, Gaussian Blur, and Solarization. Architecture.The M-IDoL pre-training framework consists of two Swin Transformer-Base (Swin-B) visual encoders, namely a student encoder Sθ and a teacher encoder Tθ. Both encoders share the same Swin-B architecture: the input im- age is split into non-overlapping 4×4 patches and linear...

  23. [23]

    baseline SwinViT (Liu et al., 2021), DINO (Caron et al.,

  24. [24]

    and MAE (He et al., 2022), and

  25. [25]

    unified pre-training MFMs Unimoss+ (Xie et al., 2024), MedCoss (Ye et al., 2024), LVM-Med (MH Nguyen et al., 2024), UniMed (Khattak et al.,

  26. [26]

    We include im- plementation code in https://github.com/LYH-hh/M-IDoL to demonstrate the reproducibility of M-IDoL

    and Tutel (Hwang et al., 2023). We include im- plementation code in https://github.com/LYH-hh/M-IDoL to demonstrate the reproducibility of M-IDoL. 17 M-IDoL: Information Decomposition for Modality-Specific and Diverse Representation Learning in Medical Foundation Model APTOS Glaucoma PAPILA Retina ISIC2016 ISIC2017 ZhangCXR RNSAHAM10000 OCTDL OCTID TVHL-D...