pith. machine review for the scientific record. sign in

arxiv: 2605.13581 · v1 · submitted 2026-05-13 · 💻 cs.CV

Recognition: unknown

HIR-ALIGN: Enhancing Hyperspectral Image Restoration via Diffusion-Based Data Generation

Authors on Pith no claims yet

Pith reviewed 2026-05-14 19:42 UTC · model grok-4.3

classification 💻 cs.CV
keywords hyperspectral image restorationdiffusion modelsdata augmentationdomain adaptationimage denoisingsuper-resolutiontarget domain finetuning
0
0 comments X

The pith

Diffusion-generated synthetic HSIs allow finetuning of restoration models to match target domains without clean references

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper establishes that source-trained hyperspectral restoration models can be adapted to new target domains by augmenting limited training data with synthetic images that closely follow the target's distribution. This is achieved through a three-stage process that starts with proxy clean HSIs generated by existing restorers, proceeds to diffusion-based RGB synthesis conditioned on those proxies, and completes with a warp-based spectral transfer to produce aligned HSIs. Finetuning on the combined proxy and synthetic data then reduces restoration risk on the target domain. Theoretical support shows the gains come from better coverage of the target distribution together with control over spectral bias. Experiments on denoising and super-resolution tasks confirm consistent improvements over source-only and unsupervised baselines.

Core claim

Augmentation-based finetuning using diffusion-generated target-aligned HSIs achieves lower target-domain restoration risk by jointly improving target distribution coverage and controlling spectral bias.

What carries the argument

The warp-based spectral transfer module, which estimates soft patch-wise transport weights from aligned RGBs and applies learnable local interpolation kernels to transfer spectra from proxy HSIs.

If this is right

  • Pretrained restoration networks show consistent gains on both simulated and real target datasets for denoising and super-resolution after aligned finetuning.
  • The combined proxy and synthetic training set improves target distribution coverage while limiting spectral bias relative to source-only training.
  • The framework remains plug-and-play and requires no extra real data beyond the degraded target observations.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The same proxy-plus-diffusion pipeline could be tested on other multi-channel imaging modalities that suffer domain shifts without paired clean data.
  • Iterating the proxy generation step with the newly finetuned model might further reduce approximation error in subsequent rounds.
  • The method's performance on real-world HSIs could be probed by varying the diffusion model's conditioning strength to isolate bias-control effects.

Load-bearing premise

Proxy HSIs from off-the-shelf restorers are semantics-preserving approximations of clean target images, and diffusion-generated RGBs can be aligned to them without introducing new biases.

What would settle it

Measuring whether finetuned restoration accuracy on real target HSIs drops to source-only levels when proxies are replaced by noisy or semantically altered versions.

Figures

Figures reproduced from arXiv: 2605.13581 by Deyu Meng, Heng Zhao, Li Pang, Xiangyong Cao, Yijia Zhang.

Figure 1
Figure 1. Figure 1: PSNR comparison of models pretrained only on ICVL [10] and the [PITH_FULL_IMAGE:figures/full_fig_p001_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: Overview of HIR-ALIGN. (a) Proxy generation: source-pretrained HSI restorers produce proxy HSIs and proxy RGBs from degraded target observations. (b) Distribution-adaptive synthesis: improved unCLIP generates target-aligned RGBs, and the proposed pixel-matching spectral transfer module maps proxy spectra to the generated RGB layout by extracting pixel features from patch descriptors, retrieving candidates … view at source ↗
Figure 3
Figure 3. Figure 3: Details of the warp based spectral transfer process. We first match the generated and proxy RGB images using multi feature patch descriptors and [PITH_FULL_IMAGE:figures/full_fig_p006_3.png] view at source ↗
Figure 4
Figure 4. Figure 4: t-SNE visualizations of (a) CLIP features derived from RGB images and (b) spectra for degraded, proxy, synthesized, and GT samples. The synthesized samples move toward the target distribution in both spatial and spectral spaces while preserving diversity [PITH_FULL_IMAGE:figures/full_fig_p008_4.png] view at source ↗
Figure 5
Figure 5. Figure 5: Examples of GT images, proxy images, and synthesized samples. The [PITH_FULL_IMAGE:figures/full_fig_p008_5.png] view at source ↗
read the original abstract

Hyperspectral image (HSI) restoration is crucial for reliable analysis, as real HSIs suffer from degradations like noise, blur, and resolution loss. However, existing models trained on source data often fail on target domains lacking clean references, a common occurrence in practice. To address this issue, we present HIR-ALIGN, a plug-and-play target-adaptive augmentation framework that enhances hyperspectral image restoration by augmenting limited training images with synthetic data that closely matches the target distribution using no extra data. It consists of three stages: (i) proxy generation, where off-the-shelf restoration models restore degraded target observations to produce semantics-preserving proxy HSIs that approximate target-domain clean images; (ii) distribution-adaptive synthesis, where a blur-robust unCLIP diffusion model generates target-aligned RGBs from proxy RGBs, with prompt conditioning and embedding-space noise initialization. Then, a warp-based spectral transfer module synthesizes HSIs by aligning each generated RGB with the proxy RGB, estimating soft patch-wise transport weights, and applying these weights and learnable local interpolation kernels to the proxy HSI; and (iii) aligned supervised finetuning, where restoration networks pretrained on the source distribution are finetuned using both the proxy HSIs and synthesized target-aligned HSIs, and are then deployed on degraded target images. We further provide theoretical analysis showing that augmentation-based finetuning can achieve lower target-domain restoration risk by jointly improving target distribution coverage and controlling spectral bias. Extensive experiments on simulated and real datasets across denoising and super-resolution tasks demonstrate that HIR-ALIGN consistently improves source-only supervised baselines, outperforming both source-only counterparts and representative unsupervised methods.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

3 major / 3 minor

Summary. The manuscript introduces HIR-ALIGN, a plug-and-play target-adaptive augmentation framework for hyperspectral image restoration. It proceeds in three stages: (i) proxy generation via off-the-shelf restorers to approximate clean target HSIs, (ii) distribution-adaptive synthesis using a blur-robust unCLIP diffusion model to generate target-aligned RGBs followed by warp-based spectral transfer (soft patch-wise transport weights and learnable local interpolation kernels) to produce aligned HSIs, and (iii) aligned supervised finetuning of source-pretrained restoration networks on the combined proxy and synthesized data. A theoretical analysis claims that this augmentation reduces target-domain restoration risk by jointly improving distribution coverage and controlling spectral bias. Experiments on simulated and real datasets report consistent gains over source-only baselines and unsupervised methods for denoising and super-resolution tasks.

Significance. If the proxy and alignment assumptions hold, the framework provides a practical route to domain adaptation for HSI restoration without requiring clean target references, which is a frequent practical constraint. The explicit theoretical link between augmentation, coverage, and bias control, together with the diffusion-based synthesis pipeline, could inform similar data-generation strategies in other imaging domains where source-target mismatch is severe.

major comments (3)
  1. [Abstract and §3.1 (proxy generation)] The central claim that augmentation-based finetuning lowers target-domain risk via improved coverage and bias control rests on the proxy HSIs being semantics-preserving approximations of clean target images. The manuscript does not isolate proxy fidelity (e.g., by direct comparison to any available target ground truth or by ablating proxy quality), leaving open whether observed gains arise from genuine distribution coverage or from incidental regularization effects of the finetuning procedure.
  2. [§3.2 (distribution-adaptive synthesis)] The warp-based spectral transfer module (soft patch-wise transport weights plus learnable local interpolation kernels) is asserted to produce unbiased aligned HSIs. Because the diffusion stage conditions on RGBs derived from potentially source-biased proxies, any semantic deviation between generated and proxy RGBs can misalign spectra; the paper provides no quantitative check (e.g., spectral angle or reconstruction error on held-out target patches) that new biases are not introduced.
  3. [Theoretical analysis section] The theoretical risk bound is presented as following from improved coverage and spectral-bias control, yet the derivation appears to rely on the proxy approximation assumption without an explicit sensitivity analysis. If the proxies retain source spectral biases or reconstruction artifacts, the coverage term in the bound may not improve as claimed.
minor comments (3)
  1. [Abstract] The abstract states that the method uses 'no extra data,' but the diffusion model is pretrained; clarify whether any target-domain images are used for prompt conditioning or embedding initialization.
  2. [§3.2] Implementation details for the learnable local interpolation kernels (e.g., kernel size, initialization, and optimization schedule) are not fully specified, making reproduction difficult.
  3. [Experiments section] Figure captions and experimental tables should report standard deviations across multiple random seeds or cross-validation folds to support the claim of 'consistent' improvement.

Simulated Author's Rebuttal

3 responses · 0 unresolved

We thank the referee for the constructive and detailed comments, which help clarify the validation of our proxy and alignment steps as well as the theoretical claims. We address each major comment below and describe the revisions planned for the manuscript.

read point-by-point responses
  1. Referee: [Abstract and §3.1 (proxy generation)] The central claim that augmentation-based finetuning lowers target-domain risk via improved coverage and bias control rests on the proxy HSIs being semantics-preserving approximations of clean target images. The manuscript does not isolate proxy fidelity (e.g., by direct comparison to any available target ground truth or by ablating proxy quality), leaving open whether observed gains arise from genuine distribution coverage or from incidental regularization effects of the finetuning procedure.

    Authors: We agree that an explicit isolation of proxy fidelity would strengthen the empirical support for our claims. In the revised manuscript we will add an ablation study on the simulated datasets (where ground truth is available) that compares restoration performance obtained with proxies from different off-the-shelf restorers against the same models trained with ground-truth clean images. We will also report proxy fidelity metrics (PSNR, SAM) on held-out simulated patches to quantify how closely the proxies approximate clean target images. These additions will help separate the contribution of improved coverage from possible regularization effects of finetuning. revision: yes

  2. Referee: [§3.2 (distribution-adaptive synthesis)] The warp-based spectral transfer module (soft patch-wise transport weights plus learnable local interpolation kernels) is asserted to produce unbiased aligned HSIs. Because the diffusion stage conditions on RGBs derived from potentially source-biased proxies, any semantic deviation between generated and proxy RGBs can misalign spectra; the paper provides no quantitative check (e.g., spectral angle or reconstruction error on held-out target patches) that new biases are not introduced.

    Authors: We acknowledge the need for quantitative verification that the warp-based transfer does not introduce additional spectral bias. In the revision we will include new experiments that compute spectral angle mapper (SAM) and per-band reconstruction error between the synthesized HSIs and the corresponding proxy HSIs on held-out target patches. We will also compare these errors against a baseline that directly uses source-biased RGBs, thereby confirming that the proposed soft transport weights and local kernels preserve spectral fidelity. revision: yes

  3. Referee: [Theoretical analysis section] The theoretical risk bound is presented as following from improved coverage and spectral-bias control, yet the derivation appears to rely on the proxy approximation assumption without an explicit sensitivity analysis. If the proxies retain source spectral biases or reconstruction artifacts, the coverage term in the bound may not improve as claimed.

    Authors: The bound is derived under the explicit modeling assumption that proxies are reasonable approximations of clean target images, as stated in the manuscript. To address the referee’s concern we will augment the theoretical section with a sensitivity analysis that quantifies how the coverage term and overall risk bound degrade under controlled levels of proxy error (e.g., additive spectral bias or reconstruction artifacts). The analysis will be accompanied by a discussion of the conditions under which the claimed improvement still holds, supported by the empirical results already showing gains across varying proxy qualities. revision: partial

Circularity Check

0 steps flagged

No circularity detected in derivation chain

full rationale

The paper describes a three-stage plug-and-play augmentation pipeline (proxy generation via off-the-shelf restorers, diffusion-based RGB synthesis with warp-based spectral transfer, and supervised finetuning) plus a separate theoretical analysis claiming lower target risk through coverage and bias control. No equations or steps in the provided text reduce a claimed prediction or first-principles result to its own fitted inputs by construction, nor do any load-bearing premises collapse to self-citations, imported uniqueness theorems, or ansatzes smuggled via prior author work. The central claims rest on stated assumptions about proxy semantics preservation rather than self-referential definitions, so the derivation chain remains independent of its outputs.

Axiom & Free-Parameter Ledger

1 free parameters · 1 axioms · 0 invented entities

The central claim rests primarily on the domain assumption that off-the-shelf models produce usable proxy approximations and that the diffusion-plus-transfer step produces distribution-matched data without new artifacts; one learnable component is introduced in the transfer module.

free parameters (1)
  • learnable local interpolation kernels
    Introduced in the warp-based spectral transfer module to apply soft patch-wise weights to the proxy HSI.
axioms (1)
  • domain assumption Off-the-shelf restoration models produce semantics-preserving proxy HSIs that approximate clean target-domain images.
    Invoked directly in the proxy generation stage as the starting point for synthesis.

pith-pipeline@v0.9.0 · 5615 in / 1368 out tokens · 70854 ms · 2026-05-14T19:42:53.316869+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

69 extracted references · 69 canonical work pages · 5 internal anchors

  1. [1]

    Coupled segmentation and denoising/deblurring models for hyperspectral material identification,

    F. Li, M. K. Ng, and R. J. Plemmons, “Coupled segmentation and denoising/deblurring models for hyperspectral material identification,” Numerical Linear Algebra with Applications, vol. 19, no. 1, pp. 153– 173, 2012

  2. [2]

    Models and methods for automated material identification in hyperspectral imagery acquired under unknown illumi- nation and atmospheric conditions,

    G. Healey and D. Slater, “Models and methods for automated material identification in hyperspectral imagery acquired under unknown illumi- nation and atmospheric conditions,”IEEE Transactions on Geoscience and Remote Sensing, vol. 37, no. 6, pp. 2706–2717, 1999

  3. [3]

    Hyperspectral image dataset for benchmarking on salient object detection,

    N. Imamoglu, Y . Oishi, X. Zhang, G. Ding, Y . Fang, T. Kouyama, and R. Nakamura, “Hyperspectral image dataset for benchmarking on salient object detection,” in2018 Tenth International Conference on Quality of Multimedia Experience (QoMEX). IEEE, 2018, pp. 1–3

  4. [4]

    Material based salient object detection from hyperspectral images,

    J. Liang, J. Zhou, L. Tong, X. Bai, and B. Wang, “Material based salient object detection from hyperspectral images,”Pattern Recognition, vol. 76, pp. 476–490, 2018

  5. [5]

    Object detection in hyperspectral images,

    L. Yan, M. Zhao, X. Wang, Y . Zhang, and J. Chen, “Object detection in hyperspectral images,”IEEE Signal Processing Letters, vol. 28, pp. 508–512, 2021

  6. [6]

    Hyperspectral imaging for clinical applications,

    J. Yoon, “Hyperspectral imaging for clinical applications,”BioChip Journal, vol. 16, no. 1, pp. 1–12, 2022

  7. [7]

    Detection of preinvasive cancer cells,

    V . Backman, M. B. Wallace, L. Perelman, J. Arendt, R. Gurjar, M. M ¨uller, Q. Zhang, G. Zonios, E. Kline, T. McGillicanet al., “Detection of preinvasive cancer cells,”Nature, vol. 406, no. 6791, pp. 35–36, 2000

  8. [8]

    Borengasser, W

    M. Borengasser, W. Hungate, and R. Watkins,Hyperspectral Remote Sensing: Principles and Applications, 12 2007

  9. [9]

    Overview of hyperspectral imaging remote sensing from satellites,

    S.-E. Qian, “Overview of hyperspectral imaging remote sensing from satellites,”Advances in Hyperspectral Image Processing Techniques, pp. 41–66, 2022

  10. [10]

    Sparse recovery of hyperspectral signal from natural rgb images,

    B. Arad and O. Ben-Shahar, “Sparse recovery of hyperspectral signal from natural rgb images,” inComputer Vision–ECCV 2016: 14th Eu- ropean Conference, Amsterdam, the Netherlands, October 11–14, 2016, Proceedings, Part VII 14. Springer, 2016, pp. 19–34. TABLE XI UPPER-BOUND COMPARISON ONCAVEANDKAISTUSINGICVL-ONLY TRAINING, HIR-ALIGN,AND CLEAN-TARGET FINETU...

  11. [11]

    Generalized assorted pixel camera: Postcapture control of resolution, dynamic range, and spectrum,

    F. Yasuma, T. Mitsunaga, D. Iso, and S. K. Nayar, “Generalized assorted pixel camera: Postcapture control of resolution, dynamic range, and spectrum,”IEEE Transactions on Image Processing, vol. 19, no. 9, pp. 2241–2253, 2010

  12. [12]

    Hyperspectral image restoration using low-rank matrix recovery,

    H. Zhang, W. He, L. Zhang, H. Shen, and Q. Yuan, “Hyperspectral image restoration using low-rank matrix recovery,”IEEE Transactions on Geoscience and Remote Sensing, vol. 52, no. 8, pp. 4729–4743, 2014

  13. [13]

    Hyperspectral image restoration using weighted group sparsity-regularized low-rank tensor decomposition,

    Y . Chen, W. He, N. Yokoya, and T.-Z. Huang, “Hyperspectral image restoration using weighted group sparsity-regularized low-rank tensor decomposition,”IEEE Transactions on Cybernetics, vol. 50, no. 8, pp. 3556–3570, 2020

  14. [14]

    Hyperspectral image denoising with weighted nonlocal low-rank model and adaptive total variation regularization,

    Y . Chen, W. Cao, L. Pang, and X. Cao, “Hyperspectral image denoising with weighted nonlocal low-rank model and adaptive total variation regularization,”IEEE Transactions on Geoscience and Remote Sensing, vol. 60, pp. 1–15, 2022

  15. [15]

    Hyperspectral image denoising via texture-preserved total variation regularizer,

    Y . Chen, W. Cao, L. Pang, J. Peng, and X. Cao, “Hyperspectral image denoising via texture-preserved total variation regularizer,”IEEE Transactions on Geoscience and Remote Sensing, vol. 61, pp. 1–14, 2023

  16. [16]

    Image restoration for remote sensing: Overview and toolbox,

    B. Rasti, Y . Chang, E. Dalsasso, L. Denis, and P. Ghamisi, “Image restoration for remote sensing: Overview and toolbox,”IEEE Geoscience and Remote Sensing Magazine, vol. 10, no. 2, pp. 201–230, 2021

  17. [17]

    Trq3dnet: A 3d quasi-recurrent and transformer based network for hyperspectral image denoising,

    L. Pang, W. Gu, and X. Cao, “Trq3dnet: A 3d quasi-recurrent and transformer based network for hyperspectral image denoising,”Remote Sensing, vol. 14, no. 18, p. 4598, 2022

  18. [18]

    Progressive hyperspectral image destriping with an adaptive frequencial focus,

    E. Pan, Y . Ma, X. Mei, F. Fan, J. Huang, and J. Ma, “Progressive hyperspectral image destriping with an adaptive frequencial focus,”IEEE Transactions on Geoscience and Remote Sensing, vol. 61, pp. 1–12, 2023

  19. [19]

    Spectral enhanced rectangle transformer for hyperspectral image denoising,

    M. Li, J. Liu, Y . Fu, Y . Zhang, and D. Dou, “Spectral enhanced rectangle transformer for hyperspectral image denoising,” inProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 5805–5814

  20. [20]

    Ssumamba: Spatial-spectral selective state space model for hyperspectral image denoising,

    G. Fu, F. Xiong, J. Lu, and J. Zhou, “Ssumamba: Spatial-spectral selective state space model for hyperspectral image denoising,”IEEE Transactions on Geoscience and Remote Sensing, vol. 62, p. 3446812,

  21. [21]

    Available: https://ui.adsabs.harvard.edu/abs/2024ITGRS

    [Online]. Available: https://ui.adsabs.harvard.edu/abs/2024ITGRS. .6246812F

  22. [22]

    Mp-hsir: A multi-prompt framework for universal hyperspectral image restoration,

    Z. Wu, Y . Chen, N. Yokoya, and W. He, “Mp-hsir: A multi-prompt framework for universal hyperspectral image restoration,”arXiv preprint arXiv:2503.09131, 2025

  23. [23]

    Denoising hyper- spectral image with non-i.i.d. noise structure,

    Y . Chen, X. Cao, Q. Zhao, D. Meng, and Z. Xu, “Denoising hyper- spectral image with non-i.i.d. noise structure,”IEEE Transactions on Cybernetics, vol. 48, no. 3, pp. 1054–1066, 2018

  24. [24]

    Hyperspectral image restoration under complex multi-band noises,

    Z. Yue, D. Meng, Y . Sun, and Q. Zhao, “Hyperspectral image restoration under complex multi-band noises,”Remote Sensing, vol. 10, no. 10, p. 1631, 2018

  25. [25]

    Hyperspectral image restoration combining intrinsic image characterization with robust noise modeling,

    T.-H. Ma, Z. Xu, D. Meng, and X.-L. Zhao, “Hyperspectral image restoration combining intrinsic image characterization with robust noise modeling,”IEEE Journal of Selected Topics in Applied Earth Observa- tions and Remote Sensing, vol. 14, pp. 1628–1644, 2021

  26. [26]

    Hyperspectral image denoising by asymmetric noise modeling,

    S. Xu, X. Cao, J. Peng, Q. Ke, C. Ma, and D. Meng, “Hyperspectral image denoising by asymmetric noise modeling,”IEEE Transactions on Geoscience and Remote Sensing, vol. 60, pp. 1–14, 2022

  27. [27]

    Adaptive hyperspectral mixed noise removal,

    T.-X. Jiang, L. Zhuang, T.-Z. Huang, and J. M. Bioucas-Dias, “Adaptive hyperspectral mixed noise removal,” inProceedings of the IEEE Inter- national Geoscience and Remote Sensing Symposium (IGARSS), 2018, pp. 4035–4038

  28. [28]

    Fasthymix: Fast and parameter-free hy- perspectral image mixed noise removal,

    L. Zhuang and M. K. Ng, “Fasthymix: Fast and parameter-free hy- perspectral image mixed noise removal,”IEEE Transactions on Neural Networks and Learning Systems, vol. 34, no. 8, pp. 4702–4716, 2023

  29. [29]

    Denoising diffusion probabilistic models,

    J. Ho, A. Jain, and P. Abbeel, “Denoising diffusion probabilistic models,” Advances in Neural Information Processing Systems, vol. 33, pp. 6840– 6851, 2020

  30. [30]

    Denoising Diffusion Implicit Models

    J. Song, C. Meng, and S. Ermon, “Denoising diffusion implicit models,” p. arXiv:2010.02502, October 01, 2020 2020, iCLR 2021; updated connections with ODEs at page 6, fixed some typos in the proof. [Online]. Available: https://ui.adsabs.harvard.edu/abs/ 2020arXiv201002502S JOURNAL OF LATEX CLASS FILES, VOL. 14, NO. 8, AUGUST 2021 15

  31. [31]

    Generative modeling by estimating gradients of the data distribution,

    Y . Song and S. Ermon, “Generative modeling by estimating gradients of the data distribution,”Advances in Neural Information Processing Systems, vol. 32, 2019

  32. [32]

    Score-Based Generative Modeling through Stochastic Differential Equations

    Y . Song, J. Sohl-Dickstein, D. P. Kingma, A. Kumar, S. Ermon, and B. Poole, “Score-based generative modeling through stochastic differential equations,” p. arXiv:2011.13456, November 01, 2020 2020, iCLR 2021 (Oral). [Online]. Available: https://ui.adsabs.harvard.edu/ abs/2020arXiv201113456S

  33. [33]

    Diffusion models beat gans on image synthesis,

    P. Dhariwal and A. Nichol, “Diffusion models beat gans on image synthesis,”Advances in Neural Information Processing Systems, vol. 34, pp. 8780–8794, 2021

  34. [34]

    GLIDE: Towards Photorealistic Image Generation and Editing with Text-Guided Diffusion Models

    A. Nichol, P. Dhariwal, A. Ramesh, P. Shyam, P. Mishkin, B. McGrew, I. Sutskever, and M. Chen, “Glide: Towards photorealistic image gen- eration and editing with text-guided diffusion models,”arXiv preprint arXiv:2112.10741, 2021

  35. [35]

    High- resolution image synthesis with latent diffusion models,

    R. Rombach, A. Blattmann, D. Lorenz, P. Esser, and B. Ommer, “High- resolution image synthesis with latent diffusion models,” inProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recogni- tion, 2022, pp. 10 684–10 695

  36. [36]

    Hierarchical Text-Conditional Image Generation with CLIP Latents

    A. Ramesh, P. Dhariwal, A. Nichol, C. Chu, and M. Chen, “Hierarchical text-conditional image generation with clip latents,” p. arXiv:2204.06125, April 01, 2022 2022. [Online]. Available: https://ui.adsabs.harvard.edu/abs/2022arXiv220406125R

  37. [37]

    Photorealistic text-to- image diffusion models with deep language understanding,

    C. Saharia, W. Chan, S. Saxena, L. Li, J. Whang, E. L. Denton, S. K. S. Ghasemipour, B. K. Ayan, S. S. Mahdavi, R. G. Lopes, T. Salimans, J. Ho, D. J. Fleet, and M. Norouzi, “Photorealistic text-to- image diffusion models with deep language understanding,” inAdvances in Neural Information Processing Systems, vol. 35, 2022, pp. 36 479– 36 494

  38. [38]

    Scalable diffusion models with transformers,

    W. Peebles and S. Xie, “Scalable diffusion models with transformers,” inProceedings of the IEEE/CVF International Conference on Computer Vision, 2023, pp. 4195–4205

  39. [39]

    Crs-diff: Controllable remote sensing image generation with diffusion model,

    D. Tang, X. Cao, X. Hou, Z. Jiang, J. Liu, and D. Meng, “Crs-diff: Controllable remote sensing image generation with diffusion model,” IEEE Transactions on Geoscience and Remote Sensing, 2024

  40. [40]

    Aerogen: enhancing remote sensing object detection with diffusion-driven data generation,

    D. Tang, X. Cao, X. Wu, J. Li, J. Yao, X. Bai, D. Jiang, Y . Li, and D. Meng, “Aerogen: enhancing remote sensing object detection with diffusion-driven data generation,” inProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2025, pp. 3614–3624

  41. [41]

    Terragen: A unified multi-task layout generation framework for remote sensing data augmentation,

    D. Tang, H. Wang, Y . Xin, H. Qiao, D. Jiang, Y . Li, Z. Yu, and X. Cao, “Terragen: A unified multi-task layout generation framework for remote sensing data augmentation,”arXiv preprint arXiv:2510.21391, 2025

  42. [42]

    Hsigene: A foundation model for hyperspectral image generation,

    L. Pang, X. Cao, D. Tang, S. Xu, X. Bai, F. Zhou, and D. Meng, “Hsigene: A foundation model for hyperspectral image generation,” arXiv preprint arXiv:2409.12470, 2024

  43. [43]

    Total-variation-regularized low-rank matrix factorization for hyperspectral image restoration,

    W. He, H. Zhang, L. Zhang, and H. Shen, “Total-variation-regularized low-rank matrix factorization for hyperspectral image restoration,”IEEE Transactions on Geoscience and Remote Sensing, vol. 54, no. 1, pp. 178–188, 2016

  44. [44]

    Hyperspectral restoration vial 0 gradient regularized low-rank tensor factorization,

    F. Xiong, J. Zhou, and Y . Qian, “Hyperspectral restoration vial 0 gradient regularized low-rank tensor factorization,”IEEE Transactions on Geoscience and Remote Sensing, vol. 57, no. 12, pp. 10 410–10 425, 2019

  45. [45]

    Enhanced 3dtv regularization and its applications on hsi denoising and compressed sensing,

    J. Peng, Q. Xie, Q. Zhao, Y . Wang, L. Yee, and D. Meng, “Enhanced 3dtv regularization and its applications on hsi denoising and compressed sensing,”IEEE Transactions on Image Processing, vol. 29, pp. 7889– 7903, 2020

  46. [46]

    Fast noise removal in hyperspectral images via representative coefficient total variation,

    J. Peng, H. Wang, X. Cao, X. Liu, X. Rui, and D. Meng, “Fast noise removal in hyperspectral images via representative coefficient total variation,”IEEE Transactions on Geoscience and Remote Sensing, vol. 60, pp. 1–17, 2022

  47. [47]

    Deep spatial-spectral global rea- soning network for hyperspectral image denoising,

    X. Cao, X. Fu, C. Xu, and D. Meng, “Deep spatial-spectral global rea- soning network for hyperspectral image denoising,”IEEE Transactions on Geoscience and Remote Sensing, vol. 60, pp. 1–14, 2022

  48. [48]

    Single hyperspectral image super-resolution with grouped deep recursive residual network,

    Y . Li, L. Zhang, C. Dingl, W. Wei, and Y . Zhang, “Single hyperspectral image super-resolution with grouped deep recursive residual network,” in2018 IEEE Fourth International Conference on Multimedia Big Data (BigMM). IEEE, 2018, pp. 1–4

  49. [49]

    Hyperspectral image superresolution using spectrum and feature context,

    Q. Wang, Q. Li, and X. Li, “Hyperspectral image superresolution using spectrum and feature context,”IEEE Transactions on Industrial Electronics, vol. 68, no. 11, pp. 11 276–11 285, 2021

  50. [50]

    Essaformer: Efficient transformer for hyperspectral image super- resolution,

    M. Zhang, C. Zhang, Q. Zhang, J. Guo, X. Gao, and J. Zhang, “Essaformer: Efficient transformer for hyperspectral image super- resolution,” in2023 IEEE/CVF International Conference on Computer Vision (ICCV), 2023, pp. 23 016–23 027

  51. [51]

    Hsr-diff: Hyperspectral image super-resolution via conditional diffusion models,

    C. Wu, D. Wang, Y . Bai, H. Mao, Y . Li, and Q. Shen, “Hsr-diff: Hyperspectral image super-resolution via conditional diffusion models,” inProceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), October 2023, pp. 7083–7093

  52. [52]

    Ispdiff: Interpretable scale- propelled diffusion model for hyperspectral image super-resolution,

    W. Dong, S. Liu, S. Xiao, J. Qu, and Y . Li, “Ispdiff: Interpretable scale- propelled diffusion model for hyperspectral image super-resolution,” IEEE Transactions on Geoscience and Remote Sensing, vol. 62, pp. 1–14, 2024

  53. [53]

    Promptir: Prompting for all-in-one image restoration,

    V . Potlapalli, S. W. Zamir, S. Khan, and F. Khan, “Promptir: Prompting for all-in-one image restoration,” inThirty-seventh Conference on Neural Information Processing Systems, 2023

  54. [54]

    Classifier-Free Diffusion Guidance

    J. Ho and T. Salimans, “Classifier-free diffusion guidance,”arXiv preprint arXiv:2207.12598, 2022

  55. [55]

    Hir-diff: Unsupervised hyperspectral image restoration via improved diffusion models,

    L. Pang, X. Rui, L. Cui, H. Wang, D. Meng, and X. Cao, “Hir-diff: Unsupervised hyperspectral image restoration via improved diffusion models,” inProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2024, pp. 3005–3014

  56. [56]

    Sdedit: guided image synthesis and editing with stochastic differential equations,

    C. Meng, Y . He, Y . Song, J. Song, J. Wu, J.-Y . Zhu, and S. Ermon, “Sdedit: guided image synthesis and editing with stochastic differential equations,” 2022

  57. [57]

    Billion-scale similarity search with gpus,

    J. Johnson, M. Douze, and H. J ´egou, “Billion-scale similarity search with gpus,”IEEE Transactions on Big Data, vol. 7, no. 3, pp. 535–547, 2019

  58. [58]

    High-quality hyperspectral reconstruction using a spectral prior,

    I. Choi, D. S. Jeon, G. Nam, D. Gutierrez, and M. H. Kim, “High-quality hyperspectral reconstruction using a spectral prior,”ACM Transactions on Graphics (Proc. SIGGRAPH Asia 2017), vol. 36, no. 6, pp. 218:1–13,

  59. [59]

    Available: http://dx.doi.org/10.1145/3130800.3130810

    [Online]. Available: http://dx.doi.org/10.1145/3130800.3130810

  60. [60]

    Hyperspectral image denoising with realistic data,

    T. Zhang, Y . Fu, and C. Li, “Hyperspectral image denoising with realistic data,” inProceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 2248–2257

  61. [61]

    A trainable spectral-spatial sparse coding model for hyperspectral image restora- tion,

    T. Bodrito, A. Zouaoui, J. Chanussot, and J. Mairal, “A trainable spectral-spatial sparse coding model for hyperspectral image restora- tion,”Advances in Neural Information Processing Systems (NeurIPS), 2021

  62. [62]

    Bidirectional 3d quasi-recurrent neural network for hyperspectral image super-resolution,

    Y . Fu, Z. Liang, and S. You, “Bidirectional 3d quasi-recurrent neural network for hyperspectral image super-resolution,”IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, vol. 14, pp. 2674–2688, 2021

  63. [63]

    Mixed 2d/3d convolutional network for hyperspectral image super-resolution,

    Q. Li, Q. Wang, and X. Li, “Mixed 2d/3d convolutional network for hyperspectral image super-resolution,”Remote Sensing, vol. 12, no. 10, 2020

  64. [64]

    Hyperspectral image restoration via total variation regularized low-rank tensor decomposition,

    Y . Wang, J. Peng, Q. Zhao, Y . Leung, X.-L. Zhao, and D. Meng, “Hyperspectral image restoration via total variation regularized low-rank tensor decomposition,”IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, vol. 11, no. 4, pp. 1227–1243, 2018

  65. [65]

    Microsoft coco: Common objects in con- text,

    T.-Y . Lin, M. Maire, S. Belongie, J. Hays, P. Perona, D. Ramanan, P. Doll´ar, and C. L. Zitnick, “Microsoft coco: Common objects in con- text,” inComputer Vision – ECCV 2014, D. Fleet, T. Pajdla, B. Schiele, and T. Tuytelaars, Eds. Cham: Springer International Publishing, 2014, pp. 740–755. JOURNAL OF LATEX CLASS FILES, VOL. 14, NO. 8, AUGUST 2021 16 Sup...

  66. [66]

    On the uniform convergence of relative frequencies of events to their probabilities,

    V . N. Vapnik and A. Y . Chervonenkis, “On the uniform convergence of relative frequencies of events to their probabilities,” inMeasures of complexity: festschrift for alexey chervonenkis. Springer, 2015, pp. 11– 30

  67. [67]

    Shalev-Shwartz and S

    S. Shalev-Shwartz and S. Ben-David,Understanding machine learning: From theory to algorithms. Cambridge university press, 2014

  68. [68]

    Mohri, A

    M. Mohri, A. Rostamizadeh, and A. Talwalkar,Foundations of machine learning. MIT press, 2018

  69. [69]

    Koltchinskii,Oracle inequalities in empirical risk minimization and sparse recovery problems: Ecole D’Et ´e de Probabilit ´es de Saint-Flour XXXVIII-2008

    V . Koltchinskii,Oracle inequalities in empirical risk minimization and sparse recovery problems: Ecole D’Et ´e de Probabilit ´es de Saint-Flour XXXVIII-2008. Springer, 2011, vol. 2033