pith. machine review for the scientific record. sign in

arxiv: 2605.14341 · v1 · submitted 2026-05-14 · 💻 cs.CV

Recognition: 2 theorem links

· Lean Theorem

AnyBand-Diff: A Unified Remote Sensing Image Generation and Band Repair Framework with Spectral Priors

Authors on Pith no claims yet

Pith reviewed 2026-05-15 01:56 UTC · model grok-4.3

classification 💻 cs.CV
keywords remote sensingdiffusion modelsspectral reconstructionimage generationphysics-guided samplingband repairspectral priors
0
0 comments X

The pith

AnyBand-Diff reconstructs complete spectral information in remote sensing images from arbitrary band subsets using physics-guided diffusion.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper introduces AnyBand-Diff to generate realistic remote sensing imagery while preserving spectral accuracy. It combines a masked conditional diffusion model with stochastic masking to handle missing bands, physics-guided sampling to enforce physical laws during denoising, and a multi-scale physical loss for consistency at different scales. This matters because standard diffusion models often produce spectral distortions that make generated data unusable for scientific Earth observation tasks. If successful, the approach allows reliable image generation and band repair without violating radiometric constraints.

Core claim

AnyBand-Diff achieves accurate spectral reconstruction and reliable imagery generation by integrating a Masked Conditional Diffusion backbone with dual stochastic masking, Physics-Guided Sampling using gradients from a differentiable physical model, and a Multi-Scale Physical Loss to enforce constraints across pixel, region, and global levels.

What carries the argument

The Physics-Guided Sampling mechanism, which leverages gradients from a differentiable physical model to steer the denoising trajectory toward physically plausible solutions.

If this is right

  • Generated images maintain radiometric fidelity suitable for quantitative analysis in remote sensing applications.
  • The framework supports reconstruction from any combination of input bands, increasing flexibility in data processing pipelines.
  • Multi-scale loss ensures consistency at local and global levels, reducing artifacts in large-scale imagery.
  • Physics integration reduces spectral distortion compared to standard diffusion approaches.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • Similar physics-guided techniques could improve generative models in other domains requiring physical consistency, such as medical imaging or climate modeling.
  • Future work might test the framework's robustness on diverse satellite sensors beyond those used in the experiments.
  • Integrating additional physical priors could further enhance performance in complex atmospheric conditions.

Load-bearing premise

The differentiable physical model accurately captures real-world radiometric and spectral relationships so that its gradients steer the process to valid solutions without new artifacts.

What would settle it

A direct comparison of the spectral signatures in generated images against independent ground-truth spectrometer measurements from the same locations would show whether the outputs match physical expectations or introduce distortions.

Figures

Figures reproduced from arXiv: 2605.14341 by Lu Li, Su Luo, Wenwen Liu, Xiaoyu Li, Ying Liu, Zuopeng Zhao.

Figure 1
Figure 1. Figure 1: The overall architecture of the AnyBand-Diff model. The input consists of a subset of spectral bands, which are randomly masked and embedded to simulate real-world sensor limitations or data loss. The model employs a PGS module to ensure spectral fidelity and physical consistency in the generated hyperspectral images. The diffusion process is carried out through Diffusion U-Net networks, with a pretraining… view at source ↗
Figure 2
Figure 2. Figure 2: Comparison of remote sensing image generation results, showcasing the capability of generating RGB and panchromatic images. bands). The loss is formulated as: Lpixel = ∥C(Xˆ 0) − S∥ 2 F , (6) where C(·) computes the Pearson correlation matrix over the spatial dimensions, and ∥ · ∥F denotes the Frobenius norm. This term prevents spectral distortion by locking the covariance structure of the synthesized band… view at source ↗
Figure 3
Figure 3. Figure 3: Quantitative comparison of AnyBand-Diff versus base￾line methods on the hyperspectral image reconstruction task under different random band masking ratios (10%, 30%, 50%). 4.2. Evaluation of Spectral Reconstruction Robustness In this experiment, we rigorously evaluate the robustness of AnyBand-Diff in reconstructing hyperspectral data under varying degrees of spectral degradation. This setup sim￾ulates rea… view at source ↗
Figure 4
Figure 4. Figure 4: Schematic of band masking. In the spectral curve compar￾ison, green points denote the masked bands. In the error heatmap, brighter colors indicate larger reconstruction errors. The comparative results on the Pavia University dataset are presented in [PITH_FULL_IMAGE:figures/full_fig_p007_4.png] view at source ↗
read the original abstract

Existing diffusion models have made significant progress in generating realistic images. However, their direct adaptation to remote sensing imagery often disregards intrinsic physical laws. This oversight frequently leads to spectral distortion and radiometric inconsistency, severely limiting the scientific utility of generated data. To address this issue, this paper introduces AnyBand-Diff, a novel spectral-prior-guided diffusion framework tailored for robust spectral reconstruction. Specifically, we design a Masked Conditional Diffusion backbone integrated with a dual stochastic masking strategy, empowering the model to recover complete spectral information from arbitrary band subsets. Subsequently, to ensure radiometric fidelity, a Physics-Guided Sampling mechanism is proposed, leveraging gradients from a differentiable physical model to explicitly steer the denoising trajectory toward the manifold of physically plausible solutions. Furthermore, a Multi-Scale Physical Loss is formulated to enforce rigorous constraints across pixel, region, and global levels in a joint manner. Extensive experiments confirm the effectiveness of AnyBand-Diff in generating reliable imagery and achieving accurate spectral reconstruction, contributing to the advancement of physics-aware generative methods for Earth observation.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

3 major / 0 minor

Summary. The paper introduces AnyBand-Diff, a diffusion-based framework for remote sensing image generation and arbitrary-band repair. It combines a Masked Conditional Diffusion backbone with dual stochastic masking to recover full spectra from partial band inputs, a Physics-Guided Sampling step that uses gradients from a differentiable physical model to steer denoising toward radiometrically plausible outputs, and a Multi-Scale Physical Loss enforcing constraints at pixel, region, and global scales. The abstract asserts that extensive experiments confirm accurate spectral reconstruction and reliable imagery generation.

Significance. If the central claims hold after proper validation, the work could advance physics-aware generative modeling for Earth observation by reducing spectral distortion and radiometric inconsistency that currently limit the scientific utility of synthetic remote-sensing data. The explicit incorporation of differentiable physical priors and multi-scale losses represents a potentially useful direction beyond standard conditional diffusion approaches.

major comments (3)
  1. [Abstract] Abstract: the claim that 'extensive experiments confirm the effectiveness' is unsupported because the abstract (and, per the provided description, the manuscript) supplies no quantitative metrics, baseline comparisons, error bars, or ablation results on held-out radiometric or spectral fidelity measures.
  2. [Physics-Guided Sampling] Physics-Guided Sampling description: the mechanism is stated to steer the denoising trajectory via gradients from a differentiable physical model, yet the manuscript provides neither the explicit equations of this model, the sensor/atmospheric/illumination parameters it employs, nor any ablation demonstrating that the steered samples improve over unguided diffusion on radiometric metrics.
  3. [Multi-Scale Physical Loss] Multi-Scale Physical Loss: the formulation is described only at a high level (pixel/region/global constraints); without the precise mathematical definition or weighting scheme, it is impossible to evaluate whether the loss actually enforces physical consistency or merely adds an auxiliary regularizer.

Simulated Author's Rebuttal

3 responses · 0 unresolved

We thank the referee for the constructive and detailed feedback. We address each major comment below and will revise the manuscript to incorporate the requested details and clarifications.

read point-by-point responses
  1. Referee: [Abstract] Abstract: the claim that 'extensive experiments confirm the effectiveness' is unsupported because the abstract (and, per the provided description, the manuscript) supplies no quantitative metrics, baseline comparisons, error bars, or ablation results on held-out radiometric or spectral fidelity measures.

    Authors: We agree that the abstract would be improved by including concrete quantitative support. In the revised manuscript we will update the abstract to summarize key results from the experiments, specifically referencing spectral reconstruction metrics (e.g., SAM and RMSE) and baseline comparisons reported in the main text and tables. revision: yes

  2. Referee: [Physics-Guided Sampling] Physics-Guided Sampling description: the mechanism is stated to steer the denoising trajectory via gradients from a differentiable physical model, yet the manuscript provides neither the explicit equations of this model, the sensor/atmospheric/illumination parameters it employs, nor any ablation demonstrating that the steered samples improve over unguided diffusion on radiometric metrics.

    Authors: The current description is high-level. We will expand Section 3.2 to include the explicit equations of the differentiable physical model, the specific sensor response functions, atmospheric parameters (e.g., from the 6S model), and illumination estimation procedure. We will also add an ablation study in the experiments section that directly compares guided versus unguided sampling on held-out radiometric and spectral fidelity metrics. revision: yes

  3. Referee: [Multi-Scale Physical Loss] Multi-Scale Physical Loss: the formulation is described only at a high level (pixel/region/global constraints); without the precise mathematical definition or weighting scheme, it is impossible to evaluate whether the loss actually enforces physical consistency or merely adds an auxiliary regularizer.

    Authors: We will revise the manuscript to provide the full mathematical definitions of the pixel-level, region-level, and global-level terms in the Multi-Scale Physical Loss, together with the exact weighting scheme used in the joint optimization. This will clarify how the loss enforces physical consistency beyond a simple regularizer. revision: yes

Circularity Check

0 steps flagged

No circularity: new architectural components and losses are independent of fitted inputs

full rationale

The derivation introduces a Masked Conditional Diffusion backbone with dual stochastic masking, Physics-Guided Sampling via gradients from a differentiable physical model, and a Multi-Scale Physical Loss. None of these reduce by construction to parameters fitted on the same data or to self-referential definitions; the physical model is invoked as an external steering mechanism rather than derived from the diffusion outputs themselves. No self-citation chains, uniqueness theorems, or renamed empirical patterns appear as load-bearing steps in the abstract or described framework. The claims rest on the proposed components plus external experiments, making the chain self-contained against benchmarks.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 0 invented entities

The central claim rests on the existence and differentiability of an accurate physical radiometric model plus standard diffusion assumptions; no free parameters or invented entities are explicitly introduced in the abstract.

axioms (1)
  • domain assumption A differentiable physical model exists that accurately captures radiometric and spectral relationships in remote sensing imagery
    Invoked directly in the Physics-Guided Sampling mechanism to steer denoising.

pith-pipeline@v0.9.0 · 5494 in / 1157 out tokens · 41753 ms · 2026-05-15T01:56:38.186233+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Lean theorems connected to this paper

Citations machine-checked in the Pith Canon. Every link opens the source theorem in the public Lean library.

What do these tags mean?
matches
The paper's claim is directly supported by a theorem in the formal canon.
supports
The theorem supports part of the paper's argument, but the paper may add assumptions or extra steps.
extends
The paper goes beyond the formal theorem; the theorem is a base layer rather than the whole result.
uses
The paper appears to rely on the theorem as machinery.
contradicts
The paper's claim conflicts with a theorem or certificate in the canon.
unclear
Pith found a possible connection, but the passage is too broad, indirect, or ambiguous to say the theorem truly supports the claim.

Reference graph

Works this paper leans on

128 extracted references · 128 canonical work pages

  1. [2]

    The changing risk and burden of wildfire in the united states

    Burke, M., Driscoll, A., Heft-Neal, S., Xue, J., Burney, J., and Wara, M. The changing risk and burden of wildfire in the united states. Proceedings of the National Academy of Sciences, 118 0 (2): 0 e2011048118, 2021

  2. [3]

    Spectraldiff: A generative framework for hyperspectral image classification with diffusion models

    Chen, N., Yue, J., Fang, L., and Xia, S. Spectraldiff: A generative framework for hyperspectral image classification with diffusion models. IEEE Transactions on Geoscience and Remote Sensing, 61: 0 1--16, 2023

  3. [4]

    Ambient diffusion: Learning clean distributions from corrupted data

    Daras, G., Shah, K., Dagan, Y., Gollakota, A., Dimakis, A., and Klivans, A. Ambient diffusion: Learning clean distributions from corrupted data. Advances in Neural Information Processing Systems, 36: 0 288--313, 2023

  4. [5]

    de Ara \'u jo, B. M. P. B., von Bloh, M., Rupprecht, V., Schaefer, H., and Asseng, S. Bird’s-eye view: Remote sensing insights into the impact of mowing events on eurasian curlew habitat selection. Agriculture, Ecosystems & Environment, 378: 0 109299, 2025

  5. [6]

    L., Xu, F., Hu, Y., B \"o sch, H., Landgraf, J., and Li, Z

    Dubovik, O., Schuster, G. L., Xu, F., Hu, Y., B \"o sch, H., Landgraf, J., and Li, Z. Grand challenges in satellite remote sensing, 2021

  6. [8]

    Scaling rectified flow transformers for high-resolution image synthesis

    Esser, P., Kulal, S., Blattmann, A., Entezari, R., M \"u ller, J., Saini, H., Levi, Y., Lorenz, D., Sauer, A., Boesel, F., et al. Scaling rectified flow transformers for high-resolution image synthesis. In Forty-first International Conference on Machine Learning, 2024

  7. [9]

    Rsvq-diffusion model for text-to-remote-sensing image generation

    Gao, X., Fu, Y., Jiang, X., Wu, F., Zhang, Y., Fu, T., Li, C., and Pei, J. Rsvq-diffusion model for text-to-remote-sensing image generation. Applied Sciences, 15 0 (3): 0 1121, 2025

  8. [10]

    M., Drees, L., Toker, A., Asseng, S., and von Bloh, M

    Goktepe, M., hossein Shamseddin, A., Uysal, E., Monteagudo, J. M., Drees, L., Toker, A., Asseng, S., and von Bloh, M. Ecomapper: Generative modeling for climate-aware satellite imagery. In Forty-second International Conference on Machine Learning, 2025

  9. [11]

    Frequency generation for real-world image super-resolution

    Guan, W., Li, H., Xu, D., Liu, J., Gong, S., and Liu, J. Frequency generation for real-world image super-resolution. IEEE Transactions on Circuits and Systems for Video Technology, 34 0 (8): 0 7029--7040, 2024

  10. [12]

    and Yu, F

    Hao, Y. and Yu, F. Super-resolution degradation model: Converting high-resolution datasets to optical zoom datasets. IEEE Transactions on Circuits and Systems for Video Technology, 33 0 (11): 0 6374--6389, 2023

  11. [13]

    Gans trained by a two time-scale update rule converge to a local nash equilibrium

    Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., and Hochreiter, S. Gans trained by a two time-scale update rule converge to a local nash equilibrium. Advances in Neural Information Processing Systems, 30, 2017

  12. [14]

    Denoising diffusion probabilistic models

    Ho, J., Jain, A., and Abbeel, P. Denoising diffusion probabilistic models. Advances in Neural Information Processing Systems, 33: 0 6840--6851, 2020

  13. [15]

    Hodson, T. O. Root mean square error (rmse) or mean absolute error (mae): When to use them or not. Geoscientific Model Development Discussions, 2022: 0 1--10, 2022

  14. [16]

    and Ziou, D

    Hore, A. and Ziou, D. Image quality metrics: Psnr vs. ssim. In 2010 20th International Conference on Pattern Recognition, pp.\ 2366--2369, 2010

  15. [17]

    and Li, T

    Hou, Y. and Li, T. Difforsinet: Salient object detection in optical remote sensing images via conditional diffusion model. IEEE Transactions on Geoscience and Remote Sensing, pp.\ 1--1, 2025

  16. [18]

    Spatiotemporal variation in land use land cover in the response to local climate change using multispectral remote sensing data

    Hussain, S., Lu, L., Mubeen, M., Nasim, W., Karuppannan, S., Fahad, S., Tariq, A., Mousa, B., Mumtaz, F., and Aslam, M. Spatiotemporal variation in land use land cover in the response to local climate change using multispectral remote sensing data. Land, 11 0 (5): 0 595, 2022

  17. [19]

    Glob-diffusion: A global consistent diffusion model for large-scale image generation

    Kang, Y., Shi, H., Liu, H., Xie, W., Fang, L., and Bruzzone, L. Glob-diffusion: A global consistent diffusion model for large-scale image generation. IEEE Transactions on Circuits and Systems for Video Technology, pp.\ 1--1, 2025

  18. [20]

    Diffusionsat: A generative foundation model for satellite imagery

    Khanna, S., Liu, P., Zhou, L., Meng, C., Rombach, R., Burke, M., Lobell, D., and Ermon, S. Diffusionsat: A generative foundation model for satellite imagery. In Kim, B., Yue, Y., Chaudhuri, S., Fragkiadaki, K., Khan, M., and Sun, Y. (eds.), International Conference on Representation Learning, volume 2024, pp.\ 5586--5604, 2024

  19. [21]

    Enhancing ship detection in remote sensing: A data augmentation approach using state-of-the-art text-to-image diffusion

    Le, T.-T.-H., Truong, T.-T.-H., and Nguyen, C.-T. Enhancing ship detection in remote sensing: A data augmentation approach using state-of-the-art text-to-image diffusion. In 2025 International Conference on Multimedia Analysis and Pattern Recognition (MAPR), pp.\ 1--6, 2025

  20. [22]

    Li, Y., Liu, H., Wu, Q., Mu, F., Yang, J., Gao, J., Li, C., and Lee, Y. J. Gligen: Open-set grounded text-to-image generation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp.\ 22511--22521, 2023

  21. [23]

    M., and Xu, B

    Li, Z., Chen, B., Wu, S., Su, M., Chen, J. M., and Xu, B. Deep learning for urban land use category classification: A review and experimental assessment. Remote Sensing of Environment, 311: 0 114290, 2024

  22. [24]

    Text2earth: Unlocking text-driven remote sensing image generation with a global-scale dataset and a foundation model

    Liu, C., Chen, K., Zhao, R., Zou, Z., and Shi, Z. Text2earth: Unlocking text-driven remote sensing image generation with a global-scale dataset and a foundation model. IEEE Geoscience and Remote Sensing Magazine, 2025 a

  23. [25]

    Black box adversarial sample generation of remote sensing image description

    Liu, G., Li, Y., Fang, S., Shang, R., and Jiao, L. Black box adversarial sample generation of remote sensing image description. In IGARSS 2025 - 2025 IEEE International Geoscience and Remote Sensing Symposium, pp.\ 6633--6636, 2025 b

  24. [26]

    Diffusion models meet remote sensing: Principles, methods, and perspectives

    Liu, Y., Yue, J., Xia, S., Ghamisi, P., Xie, W., and Fang, L. Diffusion models meet remote sensing: Principles, methods, and perspectives. IEEE Transactions on Geoscience and Remote Sensing, 2024

  25. [27]

    Ctigen-cdm: Controlled text-to-image generation using cropped diffusion models

    Liu, Y., Huang, J., Wen, S., He, X., Zhang, W., and Feng, Z. Ctigen-cdm: Controlled text-to-image generation using cropped diffusion models. IEEE Transactions on Circuits and Systems for Video Technology, 35 0 (12): 0 11849--11862, 2025 c

  26. [28]

    Y., Zhu, X

    Long, Y., Xia, G.-S., Li, S., Yang, W., Yang, M. Y., Zhu, X. X., Zhang, L., and Li, D. On creating benchmark dataset for aerial image interpretation: Reviews, guidances, and million-aid. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 14: 0 4205--4230, 2021

  27. [29]

    Exploring models and data for remote sensing image caption generation

    Lu, X., Wang, B., Zheng, X., and Li, X. Exploring models and data for remote sensing image caption generation. IEEE Transactions on Geoscience and Remote Sensing, 56 0 (4): 0 2183--2195, 2017

  28. [30]

    Large-factor super-resolution of remote sensing images with spectra-guided generative adversarial networks

    Meng, Y., Li, W., Lei, S., Zou, Z., and Shi, Z. Large-factor super-resolution of remote sensing images with spectra-guided generative adversarial networks. IEEE Transactions on Geoscience and Remote Sensing, 60: 0 1--11, 2022

  29. [31]

    T2i-adapter: Learning adapters to dig out more controllable ability for text-to-image diffusion models

    Mou, C., Wang, X., Xie, L., Wu, Y., Zhang, J., Qi, Z., and Shan, Y. T2i-adapter: Learning adapters to dig out more controllable ability for text-to-image diffusion models. In Proceedings of the AAAI Conference On Artificial Intelligence, volume 38, pp.\ 4296--4304, 2024

  30. [32]

    Physics-based generative adversarial models for image restoration and beyond

    Pan, J., Dong, J., Liu, Y., Zhang, J., Ren, J., Tang, J., Tai, Y.-W., and Yang, M.-H. Physics-based generative adversarial models for image restoration and beyond. IEEE transactions on pattern analysis and machine intelligence, 43 0 (7): 0 2449--2462, 2020

  31. [33]

    Earthsynth: Generating informative earth observation with diffusion models, 2025

    Pan, J., Lei, S., Fu, Y., Li, J., Liu, Y., Sun, Y., He, X., Peng, L., Huang, X., and Zhao, B. Earthsynth: Generating informative earth observation with diffusion models, 2025

  32. [34]

    Hsigene: A foundation model for hyperspectral image generation

    Pang, L., Cao, X., Tang, D., Xu, S., Bai, X., Zhou, F., and Meng, D. Hsigene: A foundation model for hyperspectral image generation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 48 0 (1): 0 730--746, 2026

  33. [35]

    Raissi, M., Perdikaris, P., and Karniadakis, G. E. Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. Journal of Computational physics, 378: 0 686--707, 2019

  34. [36]

    High-resolution image synthesis with latent diffusion models

    Rombach, R., Blattmann, A., Lorenz, D., Esser, P., and Ommer, B. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.\ 10684--10695, 2022

  35. [37]

    Geosynth: Contextually-aware high-resolution satellite image synthesis

    Sastry, S., Khanal, S., Dhakal, A., and Jacobs, N. Geosynth: Contextually-aware high-resolution satellite image synthesis. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.\ 460--470, 2024

  36. [38]

    and ElHelw, M

    Sebaq, A. and ElHelw, M. Rsdiff: Remote sensing image generation from text using diffusion model. Neural Computing and Applications, 36 0 (36): 0 23103--23111, 2024

  37. [39]

    D., and Chandra, R

    Shirmard, H., Farahbakhsh, E., M \"u ller, R. D., and Chandra, R. A review of machine learning in processing remote sensing data for mineral exploration. Remote Sensing of Environment, 268: 0 112750, 2022

  38. [40]

    Denoising diffusion probabilistic model with adversarial learning for remote sensing super-resolution

    Sui, J., Wu, Q., and Pun, M.-O. Denoising diffusion probabilistic model with adversarial learning for remote sensing super-resolution. Remote Sensing, 16 0 (7): 0 1219, 2024

  39. [41]

    D., Weddell, A

    Tang, C., Powell, J., Koch, D., Mullins, R. D., Weddell, A. S., and Chauhan, J. Physwin: An efficient and physically-informed foundation model for multispectral earth observation. In Advances in Neural Information Processing Systems, 2020

  40. [42]

    Crs-diff: Controllable remote sensing image generation with diffusion model

    Tang, D., Cao, X., Hou, X., Jiang, Z., Liu, J., and Meng, D. Crs-diff: Controllable remote sensing image generation with diffusion model. IEEE Transactions on Geoscience and Remote Sensing, 62: 0 1--14, 2024

  41. [43]

    Aerogen: Enhancing remote sensing object detection with diffusion-driven data generation

    Tang, D., Cao, X., Wu, X., Li, J., Yao, J., Bai, X., Jiang, D., Li, Y., and Meng, D. Aerogen: Enhancing remote sensing object detection with diffusion-driven data generation. In Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR), pp.\ 3614--3624, June 2025

  42. [44]

    Mapgen-diff: An end-to-end remote sensing image to map generator via denoising diffusion bridge model

    Tian, J., Wu, J., Chen, H., and Ma, M. Mapgen-diff: An end-to-end remote sensing image to map generator via denoising diffusion bridge model. Remote Sensing, 16 0 (19): 0 3716, 2024

  43. [45]

    and Sun, W

    Wang, C. and Sun, W. Controllable reference-based real-world remote sensing image super-resolution with generative diffusion priors, 2025

  44. [46]

    High-resolution image synthesis and semantic manipulation with conditional gans

    Wang, T.-C., Liu, M.-Y., Zhu, J.-Y., Tao, A., Kautz, J., and Catanzaro, B. High-resolution image synthesis and semantic manipulation with conditional gans. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp.\ 8798--8807, 2018

  45. [47]

    C., Sheikh, H

    Wang, Z., Bovik, A. C., Sheikh, H. R., and Simoncelli, E. P. Image quality assessment: from error visibility to structural similarity. IEEE transactions on image processing, 13 0 (4): 0 600--612, 2004

  46. [48]

    Willmott, C. J. and Matsuura, K. Advantages of the mean absolute error (mae) over the root mean square error (rmse) in assessing average model performance. Climate research, 30 0 (1): 0 79--82, 2005

  47. [49]

    Sfhn: Spatial-frequency domain hybrid network for image super-resolution

    Wu, Z., Liu, W., Li, J., Xu, C., and Huang, D. Sfhn: Spatial-frequency domain hybrid network for image super-resolution. IEEE Transactions on Circuits and Systems for Video Technology, 33 0 (11): 0 6459--6473, 2023

  48. [50]

    Data augmentation for remote sensing semantic segmentation via controllable diffusion models

    Xie, M., Gong, J., Gao, Z., and Cao, M. Data augmentation for remote sensing semantic segmentation via controllable diffusion models. In IGARSS 2025 - 2025 IEEE International Geoscience and Remote Sensing Symposium, pp.\ 6132--6136, 2025

  49. [51]

    Dldc: A dual loop data cleaning method for fine-tuning remote sensing image generative models

    Xing, T., Yan, H., Wang, X., Sun, K., Yu, H., Li, P., and Zhao, Q. Dldc: A dual loop data cleaning method for fine-tuning remote sensing image generative models. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 18: 0 28709--28725, 2025

  50. [52]

    Metaearth: A generative foundation model for global-scale remote sensing image generation

    Yu, Z., Liu, C., Liu, L., Shi, Z., and Zou, Z. Metaearth: A generative foundation model for global-scale remote sensing image generation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 47 0 (3): 0 1764--1781, 2025 a

  51. [53]

    A guideline of u-net-based framework for precipitation estimates

    Yu, Z., Wang, H., and Chen, H. A guideline of u-net-based framework for precipitation estimates. International Journal of Artificial Intelligence for Science (IJAI4S), 1 0 (1), 2025 b

  52. [54]

    J., and Asseng, S

    Zachow, M., Kunstmann, H., Miralles, D. J., and Asseng, S. Multi-model ensembles for regional and national wheat yield forecasts in argentina. Environmental Research Letters, 19 0 (8): 0 084037, 2024

  53. [55]

    Zhang, H., Xu, T., Li, H., Zhang, S., Wang, X., Huang, X., and Metaxas, D. N. Stackgan: Text to photo-realistic image synthesis with stacked generative adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, pp.\ 5907--5915, 2017

  54. [56]

    Adding conditional control to text-to-image diffusion models

    Zhang, L., Rao, A., and Agrawala, M. Adding conditional control to text-to-image diffusion models. In Proceedings of the IEEE International Conference on Computer Vision, pp.\ 3836--3847, 2023 a

  55. [57]

    Cc-diff++: Spatially controllable text-to-image synthesis for remote sensing with enhanced contextual coherence

    Zhang, M., Liu, Y., Liu, Y., Zhao, Y., and Ye, Q. Cc-diff++: Spatially controllable text-to-image synthesis for remote sensing with enhanced contextual coherence. IEEE Transactions on Geoscience and Remote Sensing, 63: 0 1--16, 2025 a

  56. [58]

    A., Shechtman, E., and Wang, O

    Zhang, R., Isola, P., Efros, A. A., Shechtman, E., and Wang, O. The unreasonable effectiveness of deep features as a perceptual metric. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp.\ 586--595, 2018

  57. [60]

    Cascaded autoregressive diffusion models for remote sensing scene generation

    Zhang, Y., Liu, L., Chen, K., Xu, J., Shi, Z., and Zou, Z. Cascaded autoregressive diffusion models for remote sensing scene generation. IEEE Transactions on Geoscience and Remote Sensing, 63: 0 1--17, 2025 b

  58. [61]

    Zhao, S., Chen, D., Chen, Y.-C., Bao, J., Hao, S., Yuan, L., and Wong, K.-Y. K. Uni-controlnet: All-in-one control to text-to-image diffusion models. Advances in Neural Information Processing Systems, 36: 0 11127--11150, 2023

  59. [62]

    Diverse text-prompt generation for remote sensing image classification

    Zhao, W., Lv, X., He, R., Zhao, F., Wang, H., and He, Y. Diverse text-prompt generation for remote sensing image classification. IEEE Transactions on Geoscience and Remote Sensing, 63: 0 1--10, 2025. doi:10.1109/TGRS.2024.3522283

  60. [63]

    Ssdiff: Spatial-spectral integrated diffusion model for remote sensing pansharpening

    Zhong, Y., Wu, X., Cao, Z., Dou, H.-X., and Deng, L.-J. Ssdiff: Spatial-spectral integrated diffusion model for remote sensing pansharpening. Advances in Neural Information Processing Systems, 37: 0 77962--77986, 2024

  61. [64]

    Zhu, J.-Y., Park, T., Isola, P., and Efros, A. A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, pp.\ 2223--2232, 2017

  62. [65]

    Diffcr: A fast conditional diffusion framework for cloud removal from optical satellite images

    Zou, X., Li, K., Xing, J., Zhang, Y., Wang, S., Jin, L., and Tao, P. Diffcr: A fast conditional diffusion framework for cloud removal from optical satellite images. IEEE Transactions on Geoscience and Remote Sensing, 62: 0 1--14, 2024

  63. [66]

    Proceedings of the IEEE International Conference on Computer Vision , pages=

    Unpaired image-to-image translation using cycle-consistent adversarial networks , author=. Proceedings of the IEEE International Conference on Computer Vision , pages=

  64. [67]

    Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition , pages=

    High-resolution image synthesis and semantic manipulation with conditional gans , author=. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition , pages=

  65. [68]

    Advances in Neural Information Processing Systems , volume=

    Denoising diffusion probabilistic models , author=. Advances in Neural Information Processing Systems , volume=

  66. [69]

    HSIGene: A Foundation Model for Hyperspectral Image Generation , year=

    Pang, Li and Cao, Xiangyong and Tang, Datao and Xu, Shuang and Bai, Xueru and Zhou, Feng and Meng, Deyu , journal=. HSIGene: A Foundation Model for Hyperspectral Image Generation , year=

  67. [70]

    IEEE Geoscience and Remote Sensing Magazine , year=

    Text2Earth: Unlocking text-driven remote sensing image generation with a global-scale dataset and a foundation model , author=. IEEE Geoscience and Remote Sensing Magazine , year=

  68. [71]

    CRS-Diff: Controllable Remote Sensing Image Generation With Diffusion Model , year=

    Tang, Datao and Cao, Xiangyong and Hou, Xingsong and Jiang, Zhongyuan and Liu, Junmin and Meng, Deyu , journal=. CRS-Diff: Controllable Remote Sensing Image Generation With Diffusion Model , year=

  69. [72]

    CC-Diff++: Spatially Controllable Text-to-Image Synthesis for Remote Sensing With Enhanced Contextual Coherence , year=

    Zhang, Mu and Liu, Yunfan and Liu, Yue and Zhao, Yuzhong and Ye, Qixiang , journal=. CC-Diff++: Spatially Controllable Text-to-Image Synthesis for Remote Sensing With Enhanced Contextual Coherence , year=

  70. [73]

    Applied Sciences , volume=

    RSVQ-Diffusion Model for Text-to-Remote-Sensing Image Generation , author=. Applied Sciences , volume=

  71. [74]

    Large-Factor Super-Resolution of Remote Sensing Images With Spectra-Guided Generative Adversarial Networks , year=

    Meng, Yapeng and Li, Wenyuan and Lei, Sen and Zou, Zhengxia and Shi, Zhenwei , journal=. Large-Factor Super-Resolution of Remote Sensing Images With Spectra-Guided Generative Adversarial Networks , year=

  72. [75]

    Journal of Computational physics , volume=

    Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations , author=. Journal of Computational physics , volume=

  73. [76]

    Advances in Neural Information Processing Systems , year=

    PhySwin: An Efficient and Physically-Informed Foundation Model for Multispectral Earth Observation , author=. Advances in Neural Information Processing Systems , year=

  74. [77]

    IEEE transactions on pattern analysis and machine intelligence , volume=

    Physics-based generative adversarial models for image restoration and beyond , author=. IEEE transactions on pattern analysis and machine intelligence , volume=

  75. [78]

    SSIM , author=

    Image quality metrics: PSNR vs. SSIM , author=. 2010 20th International Conference on Pattern Recognition , pages=

  76. [79]

    Geoscientific Model Development Discussions , volume=

    Root mean square error (RMSE) or mean absolute error (MAE): When to use them or not , author=. Geoscientific Model Development Discussions , volume=

  77. [80]

    Proceedings of the IEEE International Conference on Computer Vision , pages=

    Adding conditional control to text-to-image diffusion models , author=. Proceedings of the IEEE International Conference on Computer Vision , pages=

  78. [81]

    Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition , pages=

    High-resolution image synthesis with latent diffusion models , author=. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition , pages=

  79. [82]

    DiffusionSat: A Generative Foundation Model for Satellite Imagery , volume =

    Khanna, Samar and Liu, Patrick and Zhou, Linqi and Meng, Chenlin and Rombach, Robin and Burke, Marshall and Lobell, David and Ermon, Stefano , booktitle =. DiffusionSat: A Generative Foundation Model for Satellite Imagery , volume =

  80. [83]

    2025 , eprint=

    Controllable Reference-Based Real-World Remote Sensing Image Super-Resolution with Generative Diffusion Priors , author=. 2025 , eprint=

Showing first 80 references.