pith. machine review for the scientific record. sign in

arxiv: 2604.05562 · v1 · submitted 2026-04-07 · 💻 cs.CV

Recognition: no theorem link

Physics-Aligned Spectral Mamba: Decoupling Semantics and Dynamics for Few-Shot Hyperspectral Target Detection

Authors on Pith no claims yet

Pith reviewed 2026-05-10 19:37 UTC · model grok-4.3

classification 💻 cs.CV
keywords hyperspectral target detectionfew-shot learningMamba adapterfrequency domainparameter-efficient adaptationspectral band continuitycross-domain generalizationtest-time adaptation
0
0 comments X

The pith

SpecMamba decouples stable semantic representations from agile spectral adaptation using a frequency-domain Mamba adapter on frozen transformers for few-shot hyperspectral target detection.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper develops SpecMamba to solve the problem of adapting deep models for identifying targets in hyperspectral images when only a few labeled examples are available. Standard fine-tuning wastes computation and overfits, while most approaches overlook the natural frequency structure and continuous band relationships in spectral data. The method freezes a transformer backbone to preserve semantic stability, then adds a lightweight adapter that transforms features into the frequency domain and applies state-space recursion to model spectral dependencies efficiently. Laboratory spectral priors guide adaptation without shifting the core features, and a self-supervised step refines boundaries at test time. If effective, this yields higher detection rates and better transfer across different sensors or environments.

Core claim

SpecMamba decouples stable semantic representation from agile spectral adaptation by introducing a Discrete Cosine Transform Mamba Adapter (DCTMA) on top of frozen Transformer representations. The DCTMA projects spectral features into the frequency domain via DCT and leverages Mamba's linear-complexity state-space recursion to capture global spectral dependencies and band continuity. A Prior-Guided Tri-Encoder (PGTE) incorporates laboratory spectral priors to guide adapter optimization, and a Self-Supervised Pseudo-Label Mapping (SSPLM) enables test-time adaptation through uncertainty-aware sampling and dual-path consistency.

What carries the argument

The Discrete Cosine Transform Mamba Adapter (DCTMA), which projects spectral features into the frequency domain via DCT and applies Mamba state-space recursion to capture global dependencies and band continuity without full fine-tuning.

If this is right

  • SpecMamba achieves higher detection accuracy than prior state-of-the-art methods across multiple public hyperspectral datasets.
  • The framework improves cross-domain generalization in few-shot regimes by keeping semantic features fixed while adapting only spectral dynamics.
  • Parameter efficiency is obtained by freezing the transformer backbone and training only the lightweight adapter plus encoders.
  • Test-time refinement via SSPLM sharpens decision boundaries using pseudo-labels derived from uncertainty sampling and consistency constraints.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The same decoupling pattern could extend to other sequential or band-structured data types, such as time-series spectroscopy or multi-band medical imaging, where stable semantics must be preserved amid variable dynamics.
  • If the frequency-domain projection proves robust, it suggests state-space models may offer advantages over attention for low-data spectral tasks because of their linear scaling with sequence length.
  • The reliance on laboratory priors raises the possibility of hybrid systems that combine measured reference spectra with image-derived features for even stronger guidance in novel domains.

Load-bearing premise

Projecting spectral features into the frequency domain via DCT and applying Mamba recursion will reliably capture global spectral dependencies and band continuity without introducing artifacts that degrade detection in real hyperspectral data.

What would settle it

A controlled test on a hyperspectral dataset containing strong band discontinuities or sensor-specific artifacts where SpecMamba shows no accuracy gain or underperforms non-frequency baselines would falsify the claim that the DCTMA reliably extracts useful dependencies.

Figures

Figures reproduced from arXiv: 2604.05562 by Chao Li, Fanda Fan, Luqi Gong, Qixin Xie, Shuai Zhao, Yue Chen, Ziqiang Chen.

Figure 1
Figure 1. Figure 1: Motivation of the proposed SpecMamba. Conventional direct spectral [PITH_FULL_IMAGE:figures/full_fig_p001_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: Overview of the proposed SpecMamba framework. It consists of three stages: data processing, dual-stream spectral-spatial meta-learning with DCTMA [PITH_FULL_IMAGE:figures/full_fig_p004_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: Detailed architecture of the selective SSM in the DCTMA module. It combines DCT-based frequency decomposition with Mamba-based state-space [PITH_FULL_IMAGE:figures/full_fig_p005_3.png] view at source ↗
Figure 4
Figure 4. Figure 4: Pseudo-color image and Ground truth of the Source data (Chikusei). [PITH_FULL_IMAGE:figures/full_fig_p006_4.png] view at source ↗
Figure 5
Figure 5. Figure 5: Dataset image scenes and ground truth. (a) San Diego I. (b) San Diego [PITH_FULL_IMAGE:figures/full_fig_p007_5.png] view at source ↗
Figure 6
Figure 6. Figure 6: Experimental datasets and visualized detection maps of competing methods. [PITH_FULL_IMAGE:figures/full_fig_p008_6.png] view at source ↗
Figure 7
Figure 7. Figure 7: ROC curves for each detection method on the four datasets, from left to right: 3-D ROC curves, 2-D ROC curves [PITH_FULL_IMAGE:figures/full_fig_p008_7.png] view at source ↗
Figure 8
Figure 8. Figure 8: Target–background separation diagrams of competing methods on four datasets. [PITH_FULL_IMAGE:figures/full_fig_p009_8.png] view at source ↗
Figure 9
Figure 9. Figure 9: Fig. S1. Sensitivity analysis of SpecMamba with respect to four key [PITH_FULL_IMAGE:figures/full_fig_p011_9.png] view at source ↗
Figure 10
Figure 10. Figure 10: Fig. S2. Sensitivity analysis of pseudo-label confidence thresholds and [PITH_FULL_IMAGE:figures/full_fig_p012_10.png] view at source ↗
read the original abstract

Meta-learning facilitates few-shot hyperspectral target detection (HTD), but adapting deep backbones remains challenging. Full-parameter fine-tuning is inefficient and prone to overfitting, and existing methods largely ignore the frequency-domain structure and spectral band continuity of hyperspectral data, limiting spectral adaptation and cross-domain generalization.To address these challenges, we propose SpecMamba, a parameter-efficient and frequency-aware framework that decouples stable semantic representation from agile spectral adaptation. Specifically, we introduce a Discrete Cosine Transform Mamba Adapter (DCTMA) on top of frozen Transformer representations. By projecting spectral features into the frequency domain via DCT and leveraging Mamba's linear-complexity state-space recursion, DCTMA explicitly captures global spectral dependencies and band continuity while avoiding the redundancy of full fine-tuning. Furthermore, to address prototype drift caused by limited sample sizes, we design a Prior-Guided Tri-Encoder (PGTE) that allows laboratory spectral priors to guide the optimization of the learnable adapter without disrupting the stable semantic feature space. Finally, a Self-Supervised Pseudo-Label Mapping (SSPLM) strategy is developed for test-time adaptation, enabling efficient decision boundary refinement through uncertainty-aware sampling and dual-path consistency constraints. Extensive experiments on multiple public datasets demonstrate that SpecMamba consistently outperforms state-of-the-art methods in detection accuracy and cross-domain generalization.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

1 major / 2 minor

Summary. The paper proposes SpecMamba, a parameter-efficient and frequency-aware framework for few-shot hyperspectral target detection (HTD). It decouples stable semantic representations (from frozen Transformer backbones) from agile spectral adaptation via three components: the Discrete Cosine Transform Mamba Adapter (DCTMA), which projects features into the frequency domain using DCT and applies Mamba state-space recursion to capture global spectral dependencies and band continuity; the Prior-Guided Tri-Encoder (PGTE), which incorporates laboratory spectral priors to mitigate prototype drift; and the Self-Supervised Pseudo-Label Mapping (SSPLM) for test-time adaptation via uncertainty-aware sampling and consistency constraints. The central claim is that this approach consistently outperforms state-of-the-art methods in detection accuracy and cross-domain generalization on multiple public datasets while avoiding the inefficiency and overfitting of full fine-tuning.

Significance. If the empirical results and ablations hold, the work has moderate significance for few-shot learning in hyperspectral imaging. It addresses practical challenges of adapting deep models to spectral data with limited labels by incorporating frequency-domain structure and linear-complexity recursion, which could improve efficiency and generalization in remote-sensing applications. The explicit use of physics-aligned priors and Mamba for spectral continuity is a constructive direction, though its advantage depends on validation against the frequency-domain assumptions.

major comments (1)
  1. [§3.2 (DCTMA)] §3.2 (DCTMA): The load-bearing premise that DCT projection plus Mamba recursion 'explicitly captures global spectral dependencies and band continuity' without introducing artifacts (phase shifts, aliasing, or loss of fine-grained absorption features) is not supported by any derivation, comparison to alternative bases (e.g., wavelets or learned filters), or ablation isolating DCT-induced distortion on real hyperspectral signatures. If this premise fails, the claimed decoupling, avoidance of full fine-tuning redundancy, and cross-domain gains collapse.
minor comments (2)
  1. The abstract and method descriptions would benefit from explicit dataset names, quantitative metrics (e.g., AUC improvements with error bars), and statistical significance tests to ground the 'consistent outperformance' claim.
  2. [§3.2] Notation for the Mamba recursion and DCT coefficients should be defined with equations in §3.2 for reproducibility.

Simulated Author's Rebuttal

1 responses · 0 unresolved

We thank the referee for the constructive comment on the DCTMA design. We address the concern directly below and agree that additional support is warranted to substantiate the frequency-domain claims.

read point-by-point responses
  1. Referee: [§3.2 (DCTMA)] §3.2 (DCTMA): The load-bearing premise that DCT projection plus Mamba recursion 'explicitly captures global spectral dependencies and band continuity' without introducing artifacts (phase shifts, aliasing, or loss of fine-grained absorption features) is not supported by any derivation, comparison to alternative bases (e.g., wavelets or learned filters), or ablation isolating DCT-induced distortion on real hyperspectral signatures. If this premise fails, the claimed decoupling, avoidance of full fine-tuning redundancy, and cross-domain gains collapse.

    Authors: We acknowledge that the current manuscript provides no formal derivation of artifact-free behavior, no direct comparison against wavelets or learned bases, and no ablation that isolates DCT-induced distortion on hyperspectral absorption lines. DCT was selected for its real-valued, energy-compacting, and invertible properties that align with the band-continuous nature of reflectance spectra; the subsequent Mamba recursion then models long-range dependencies across the resulting coefficients at linear cost. Nevertheless, these design choices remain empirically motivated rather than theoretically proven within the paper. To close this gap we will (i) add a short paragraph in §3.2 deriving the invertibility and continuity-preserving properties of DCT for band-limited signals, (ii) expand the ablation study with wavelet and learned-filter baselines, and (iii) include signature-reconstruction visualizations that quantify preservation of fine absorption features before and after the DCTMA block. These revisions will appear in the main text and supplementary material. revision: yes

Circularity Check

0 steps flagged

No significant circularity; claims rest on architectural design using standard transforms and external priors

full rationale

The paper proposes SpecMamba with DCTMA (DCT projection + Mamba recursion on frozen Transformers), PGTE (lab priors guiding adapters), and SSPLM (self-supervised test-time adaptation). These are presented as new modules that decouple semantics from spectral adaptation, with no equations shown that define a quantity in terms of itself, fit a parameter to data then rename the fit as a prediction, or rely on self-citations for uniqueness theorems. The abstract and described components use standard DCT and Mamba as building blocks without tautological reduction; performance claims are evaluated on public datasets rather than forced by internal definitions. This qualifies as self-contained with independent content.

Axiom & Free-Parameter Ledger

0 free parameters · 0 axioms · 0 invented entities

Abstract-only; no explicit free parameters, axioms, or invented physical entities are stated. The framework relies on standard assumptions about Transformer representations and DCT invertibility.

pith-pipeline@v0.9.0 · 5559 in / 1152 out tokens · 30072 ms · 2026-05-10T19:37:39.233648+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

66 extracted references · 4 canonical work pages · 2 internal anchors

  1. [1]

    Collaborative-guided spectral abundance learning with bilinear mixing model for hyperspectral subpixel target detection,

    D. Zhu, B. Du, M. Hu, Y. Dong, and L. Zhang, “Collaborative-guided spectral abundance learning with bilinear mixing model for hyperspectral subpixel target detection,”Neural Networks, vol. 163, pp. 205–218, 2023

  2. [2]

    Hyperspectral tar- get detection with target prior augmentation and background suppression- based multidetector fusion,

    T. Guo, F. Luo, J. Guo, Y. Duan, X. Huang, and G. Shi, “Hyperspectral tar- get detection with target prior augmentation and background suppression- based multidetector fusion,”IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, vol. 17, pp. 1765–1780, 2024

  3. [3]

    Guided real image dehazing using ycbcr color space,

    W. Fang, J. Fan, Y. Zheng, J. Weng, Y. Tai, and J. Li, “Guided real image dehazing using ycbcr color space,” inProceedings of the AAAI conference on artificial intelligence, vol. 39, no. 3, 2025, pp. 2906–2914

  4. [4]

    Advances in hyperspectral image and signal processing: A comprehensive overview of the state of the art,

    P. Ghamisi, N. Yokoya, J. Li, W. Liao, S. Liu, J. Plaza, B. Rasti, and A. Plaza, “Advances in hyperspectral image and signal processing: A comprehensive overview of the state of the art,”IEEE Geoscience and Remote Sensing Magazine, vol. 5, no. 4, pp. 37–78, 2017

  5. [5]

    Meta-learning based hyperspectral target detection using siamese network,

    Y. Wang, X. Chen, F. Wang, M. Song, and C. Yu, “Meta-learning based hyperspectral target detection using siamese network,”IEEE Transactions on Geoscience and Remote Sensing, vol. 60, p. 1–13, 2022. [Online]. Available: http://dx.doi.org/10.1109/tgrs.2022.3169970

  6. [6]

    Deep spatial–spectral joint-sparse prior encoding network for hyperspectral target detection,

    W. Dong, X. Wu, J. Qu, P. Gamba, S. Xiao, A. Vizziello, and Y. Li, “Deep spatial–spectral joint-sparse prior encoding network for hyperspectral target detection,” Jun 2024

  7. [7]

    Data- augmented matched subspace detector for hyperspectral subpixel target detection,

    X. Yang, M. Dong, Z. Wang, L. Gao, L. Zhang, and J.-H. Xue, “Data- augmented matched subspace detector for hyperspectral subpixel target detection,”Pattern Recognition, vol. 106, p. 107464, 2020

  8. [8]

    Hyperspectral remote sensing image subpixel target detection based on supervised metric learning,

    L. Zhang, L. Zhang, D. Tao, X. Huang, and B. Du, “Hyperspectral remote sensing image subpixel target detection based on supervised metric learning,”IEEE Transactions on Geoscience and Remote Sensing, vol. 52, no. 8, pp. 4955–4965, 2014

  9. [9]

    Target detection with spatial- spectral adaptive sample generation and deep metric learning for hyper- spectral imagery,

    D. Zhu, B. Du, Y. Dong, and L. Zhang, “Target detection with spatial- spectral adaptive sample generation and deep metric learning for hyper- spectral imagery,”IEEE Transactions on Multimedia, vol. 25, pp. 6538– 6550, 2023

  10. [10]

    Hyperspectral target detection based on interpretable representation network,

    D. Shen, X. Ma, W. Kong, J. Liu, J. Wang, and H. Wang, “Hyperspectral target detection based on interpretable representation network,”IEEE Transactions on Geoscience and Remote Sensing, vol. 61, pp. 1–16, 2023

  11. [11]

    Exploring an application-oriented land-based hyperspectral target detection framework based on 3d–2d cnn and transfer learning,

    J. Zhao, G. Wang, B. Zhou, J. Ying, and J. Liu, “Exploring an application-oriented land-based hyperspectral target detection framework based on 3d–2d cnn and transfer learning,”EURASIP Journal on Advances in Signal Processing, vol. 2024, no. 1, Mar 2024. [Online]. Available: http://dx.doi.org/10.1186/s13634-024-01136-0

  12. [12]

    Triplet spectralwise transformer network for hyperspectral target detection,

    J. Jiao, Z. Gong, and P. Zhong, “Triplet spectralwise transformer network for hyperspectral target detection,”IEEE Trans. Geosci. Remote Sens., vol. 61, Art. no. 5519817, 2023

  13. [13]

    Spectral discrepancy and cross-modal semantic consistency learning for object detection in hyperspectral images,

    X. He, C. Tang, X. Liu, W. Zhang, Z. Gao, C. Li, S. Qiu, and J. Xu, “Spectral discrepancy and cross-modal semantic consistency learning for object detection in hyperspectral images,”IEEE Transactions on Multimedia, 2025

  14. [14]

    Htd-ts3: Weakly supervised hyperspec- tral target detection based on transformer via spectral–spatial similarity,

    H. Qin, W. Xie, Y. Li, and Q. Du, “Htd-ts3: Weakly supervised hyperspec- tral target detection based on transformer via spectral–spatial similarity,” IEEE Transactions on Neural Networks and Learning Systems, vol. 35, no. 11, pp. 16 816–16 830, 2024

  15. [15]

    Semantic modeling of hyperspectral target detection with weak labels,

    C. Jiao, B. Yang, L. Liu, C. Chen, X. Chen, W. Yang, and L. Jiao, “Semantic modeling of hyperspectral target detection with weak labels,” Signal Processing, vol. 209, p. 109016, 2023

  16. [16]

    Transfer learning of spatial features from high-resolution rgb images for large-scale and robust hyperspectral remote sensing target detection,

    Y. Wu, Z. Li, B. Zhao, Y. Song, and B. Zhang, “Transfer learning of spatial features from high-resolution rgb images for large-scale and robust hyperspectral remote sensing target detection,”IEEE Transactions on Geoscience and Remote Sensing, vol. 62, pp. 1–32, 2024

  17. [17]

    Proxy-enhanced prototype memory network for weakly supervised hyperspectral target detection,

    B. Yang, J. Wu, and C. Jiao, “Proxy-enhanced prototype memory network for weakly supervised hyperspectral target detection,”IEEE Transactions on Geoscience and Remote Sensing, vol. 63, pp. 1–14, 2025

  18. [18]

    Spatial–frequency domain transformation for infrared small target detection,

    Y. Liu, B. Tu, B. Liu, Y. He, J. Li, and A. Plaza, “Spatial–frequency domain transformation for infrared small target detection,”IEEE Transactions on Geoscience and Remote Sensing, vol. 63, pp. 1–16, 2025

  19. [19]

    Hyperspectral time-series target detection based on spectral perception and spatial–temporal tensor decomposition,

    X. Zhao, K. Liu, K. Gao, and W. Li, “Hyperspectral time-series target detection based on spectral perception and spatial–temporal tensor decomposition,”IEEE Transactions on Geoscience and Remote Sensing, vol. 61, pp. 1–12, 2023

  20. [20]

    Spectrum-driven 10 mixed-frequency network for hyperspectral salient object detection,

    P. Liu, T. Xu, H. Chen, S. Zhou, H. Qin, and J. Li, “Spectrum-driven 10 mixed-frequency network for hyperspectral salient object detection,”IEEE Transactions on Multimedia, vol. 26, pp. 5296–5310, 2024

  21. [21]

    The spectral image processing system (SIPS)—interactive visualization and analysis of imaging spectrometer data,

    F. A. Kruse, A. Lefkoff, y. J. Boardman, K. Heidebrecht, A. Shapiro, P. Barloon, and A. Goetz, “The spectral image processing system (SIPS)—interactive visualization and analysis of imaging spectrometer data,”Remote Sens. Environ., vol. 44, no. 2, pp. 145–163, 1993

  22. [22]

    An information-theoretic approach to spectral variability, similarity, and discrimination for hyperspectral image analysis,

    C.-I. Chang, “An information-theoretic approach to spectral variability, similarity, and discrimination for hyperspectral image analysis,”IEEE Trans. Inf. Theory, vol. 46, no. 5, pp. 1927–1932, 2000

  23. [23]

    The cfar adaptive subspace detector is a scale- invariant glrt,

    S. Kraut and L. Scharf, “The cfar adaptive subspace detector is a scale- invariant glrt,”IEEE Trans. Signal Process., vol. 47, no. 9, pp. 2538–2541, 1999

  24. [24]

    Hyperspectral image classification and dimensionality reduction: an orthogonal subspace projection approach,

    J. Harsanyi and C.-I. Chang, “Hyperspectral image classification and dimensionality reduction: an orthogonal subspace projection approach,” IEEE Trans. Geosci. Remote Sens., vol. 32, no. 4, pp. 779–785, 1994

  25. [25]

    Orthogonal subspace projection (osp) revisited: A compre- hensive study and analysis,

    C.-I. Chang, “Orthogonal subspace projection (osp) revisited: A compre- hensive study and analysis,”IEEE Trans. Geosci. Remote Sens., vol. 43, no. 3, pp. 502–518, 2005

  26. [26]

    A comparative study for orthogonal subspace projection and constrained energy minimization,

    Q. Du, H. Ren, and C.-I. Chang, “A comparative study for orthogonal subspace projection and constrained energy minimization,”IEEE Trans. Geosci. Remote Sens., vol. 41, no. 6, pp. 1525–1529, 2003

  27. [27]

    Constrained subpixel target detection for remotely sensed imagery,

    C.-I. Chang and D. Heinz, “Constrained subpixel target detection for remotely sensed imagery,”IEEE Trans. Geosci. Remote Sens., vol. 38, no. 3, pp. 1144–1159, 2000

  28. [28]

    Hierarchical suppression method for hyperspectral target detection,

    Z. Zou and Z. Shi, “Hierarchical suppression method for hyperspectral target detection,”IEEE Trans. Geosci. Remote Sens., vol. 54, no. 1, pp. 330–342, 2015

  29. [29]

    Ensemble-based cascaded constrained energy minimization for hyperspectral target detection,

    R. Zhao, Z. Shi, Z. Zou, and Z. Zhang, “Ensemble-based cascaded constrained energy minimization for hyperspectral target detection,” Remote Sens., vol. 11, no. 11, p. 1310, 2019

  30. [30]

    Kernel matched subspace detectors for hyperspectral target detection,

    H. Kwon and N. M. Nasrabadi, “Kernel matched subspace detectors for hyperspectral target detection,”IEEE Trans. Pattern Anal. Mach. Intell., vol. 28, no. 2, pp. 178–194, 2006

  31. [31]

    Combined sparse and collaborative representation for hyperspectral target detection,

    W. Li, Q. Du, and B. Zhang, “Combined sparse and collaborative representation for hyperspectral target detection,”Pattern Recognition, vol. 48, no. 12, pp. 3904–3916, 2015

  32. [32]

    A hybrid sparsity and distance-based discrimination detector for hyperspectral images,

    X. Lu, W. Zhang, and X. Li, “A hybrid sparsity and distance-based discrimination detector for hyperspectral images,”IEEE Transactions on Geoscience and Remote Sensing, vol. 56, no. 3, pp. 1704–1717, 2018

  33. [33]

    Learning tensor low-rank representation for hyperspectral anomaly detection,

    M. Wang, Q. Wang, D. Hong, S. K. Roy, and J. Chanussot, “Learning tensor low-rank representation for hyperspectral anomaly detection,”IEEE Transactions on Cybernetics, vol. 53, no. 1, pp. 679–691, 2023

  34. [34]

    Hyperspectral target detection based on generative self-supervised learning with wavelet transform,

    H. Qin, S. Wang, Y. Li, W. Xie, K. Jiang, and K. Cao, “Hyperspectral target detection based on generative self-supervised learning with wavelet transform,”IEEE Transactions on Geoscience and Remote Sensing, vol. 63, pp. 1–15, 2025

  35. [35]

    Hyperspectral target detection using diffusion model and convolutional gated linear unit,

    Z. Li, F. Mo, X. Zhao, C. Liu, X. Gong, W. Li, Q. Du, and B. Yuan, “Hyperspectral target detection using diffusion model and convolutional gated linear unit,”IEEE Transactions on Geoscience and Remote Sensing, vol. 63, pp. 1–13, 2025

  36. [36]

    A signature- constrained two-stage framework for hyperspectral target detection based on generative self-supervised learning,

    H. Qin, S. Wang, Y. Li, W. Xie, K. Jiang, and K. Cao, “A signature- constrained two-stage framework for hyperspectral target detection based on generative self-supervised learning,”IEEE Transactions on Geo- science and Remote Sensing, vol. 63, pp. 1–17, 2025

  37. [37]

    Transferred deep learning for hyperspectral target detection,

    W. Li, G. Wu, and Q. Du, “Transferred deep learning for hyperspectral target detection,” inProc. IEEE Int. Geosci. Remote Sens. Symp., 2017, pp. 5177–5180

  38. [38]

    HTD-Net: A deep convolutional neural network for target detection in hyperspectral imagery,

    G. Zhang, S. Zhao, W. Li, Q. Du, Q. Ran, and R. Tao, “HTD-Net: A deep convolutional neural network for target detection in hyperspectral imagery,”Remote Sens., vol. 12, no. 9, 2020

  39. [39]

    Hyperspectral target detection based on interpretable representation network,

    D. Shen, X. Ma, W. Kong, J. Liu, J. Wang, and H. Wang, “Hyperspectral target detection based on interpretable representation network,”IEEE Trans. Geosci. Remote Sens., vol. 61, pp. 1–16, 2023

  40. [40]

    Htd-net: A deep convolutional neural network for target detection in hyperspectral imagery,

    G. Zhang, S. Zhao, W. Li, Q. Du, Q. Ran, and R. Tao, “Htd-net: A deep convolutional neural network for target detection in hyperspectral imagery,”Remote Sensing, vol. 12, no. 9, 2020

  41. [41]

    Siamese transformer network for hyperspectral image target detection,

    W. Rao, L. Gao, Y. Qu, X. Sun, B. Zhang, and J. Chanussot, “Siamese transformer network for hyperspectral image target detection,”IEEE Trans. Geosci. Remote Sens., vol. 60, Art. no. 5526419, 2022

  42. [42]

    Htdformer: Hyperspectral target detection based on transformer with distributed learning,

    Y. Li, H. Qin, and W. Xie, “Htdformer: Hyperspectral target detection based on transformer with distributed learning,”IEEE Trans. Geosci. Remote Sens., vol. 61, 2023

  43. [43]

    Self-supervised spectral- level contrastive learning for hyperspectral target detection,

    Y. Wang, X. Chen, E. Zhao, and M. Song, “Self-supervised spectral- level contrastive learning for hyperspectral target detection,”IEEE Trans. Geosci. Remote Sens., vol. 61, Art. no. 5510515, 2023

  44. [44]

    Deformable convolution- enhanced hierarchical transformer with spectral-spatial cluster attention for hyperspectral image classification,

    Y. Fang, L. Sun, Y. Zheng, and Z. Wu, “Deformable convolution- enhanced hierarchical transformer with spectral-spatial cluster attention for hyperspectral image classification,”IEEE Trans. Image Process., vol. 34, pp. 701–716, 2025

  45. [45]

    Deep- growing neural network with manifold constraints for hyperspectral image classification,

    J. Shi, T. Wu, A. K. Qin, T. Shao, Y. Lei, and G. Jeon, “Deep- growing neural network with manifold constraints for hyperspectral image classification,”IEEE Trans. Neural Netw. Learn. Syst., vol. 36, no. 1, pp. 210–221, 2025

  46. [46]

    An unsupervised mo- mentum contrastive learning based transformer network for hyperspectral target detection,

    Y. Wang, X. Chen, E. Zhao, M. Song, and C. Yu, “An unsupervised mo- mentum contrastive learning based transformer network for hyperspectral target detection,”Remote Sens., vol. 17, no. 19, pp. 9053–9068, 2024

  47. [47]

    Adaptive temperature- driven ternary contrastive autoencoder framework for hyperspectral target detection,

    J. Li, H. Liu, H. Xu, C. Wang, Y. Li, and Q. Du, “Adaptive temperature- driven ternary contrastive autoencoder framework for hyperspectral target detection,”IEEE Transactions on Geoscience and Remote Sensing, vol. 63, pp. 1–13, 2025

  48. [48]

    Sensor-independent hyperspectral target detection with semisupervised domain adaptive few-shot learning,

    Y. Shi, J. Li, Y. Li, and Q. Du, “Sensor-independent hyperspectral target detection with semisupervised domain adaptive few-shot learning,”IEEE Trans. Geosci. Remote Sens., vol. 59, no. 8, pp. 6894–6906, 2021

  49. [49]

    Meta-learning based hyperspectral target detection using siamese network,

    Y. Wang, X. Chen, Y. Fang, M. Song, and C. Yu, “Meta-learning based hyperspectral target detection using siamese network,”IEEE Trans. Geosci. Remote Sens., vol. 60, 2022

  50. [50]

    Regularized tensor representative coefficient model for hyperspectral target detection,

    W. Shang, M. Jouni, Z. Wu, Y. Xu, M. D. Mura, and Z. Wei, “Regularized tensor representative coefficient model for hyperspectral target detection,” IEEE Geoscience and Remote Sensing Letters, vol. 20, pp. 1–5, 2023

  51. [51]

    Background learning based on target suppression constraint for hyperspectral target detection,

    W. Xie, X. Zhang, Y. Yang, K. Li, W. Wang, and Q. Du, “Background learning based on target suppression constraint for hyperspectral target detection,”IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens., vol. 13, pp. 5887–5897, 2021

  52. [52]

    Transformer-based cross-domain few-shot learning for hyperspectral target detection,

    S. Feng, X. Wang, R. Feng, F. Xiong, C. Zhao, W. Li, and R. Tao, “Transformer-based cross-domain few-shot learning for hyperspectral target detection,”IEEE Transactions on Geoscience and Remote Sensing, vol. 63, pp. 1–16, 2025

  53. [53]

    Unbalanced episode meta-learning with bi-sparse contrastive network for hyperspectral target detection,

    Q. Liu, Y. Xu, Z. Wu, J. Peng, and Z. Wei, “Unbalanced episode meta-learning with bi-sparse contrastive network for hyperspectral target detection,”Pattern Recognition, vol. 170, p. 112030, 2026

  54. [54]

    Mamba: Linear-Time Sequence Modeling with Selective State Spaces

    A. Gu and T. Dao, “Mamba: Linear-time sequence modeling with selective state spaces,”arXiv preprint arXiv:2312.00752, 2023

  55. [55]

    Efficiently Modeling Long Sequences with Structured State Spaces

    A. Gu, K. Goel, and C. Re, “Efficiently modeling long sequences with structured state spaces,”arXiv preprint arXiv:2111.00396, 2021

  56. [56]

    Sedgm: A structure-enhanced spatial–spectral dynamic gating mamba for hyperspectral image classification,

    Y. Jiang, S. Zhang, C. Wang, G. Zhang, M. Tan, B. Du, and X. Shen, “Sedgm: A structure-enhanced spatial–spectral dynamic gating mamba for hyperspectral image classification,”IEEE Transactions on Geoscience and Remote Sensing, vol. 63, pp. 1–22, 2025

  57. [57]

    3dss-mamba: 3d-spectral- spatial mamba for hyperspectral image classification,

    Y. He, B. Tu, B. Liu, J. Li, and A. Plaza, “3dss-mamba: 3d-spectral- spatial mamba for hyperspectral image classification,”IEEE Transactions on Geoscience and Remote Sensing, 2024

  58. [58]

    Spectral–spatial feature extraction network with ssm–cnn for hyperspectral–multispectral image collaborative classification,

    Q. Wang, X. Fan, J. Huang, S. Li, and T. Shen, “Spectral–spatial feature extraction network with ssm–cnn for hyperspectral–multispectral image collaborative classification,”IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 2024

  59. [59]

    Meta-learning of neural state-space models using data from similar systems,

    A. Chakrabarty, G. Wichern, and C. R. Laughman, “Meta-learning of neural state-space models using data from similar systems,”IFAC- PapersOnLine, vol. 56, no. 2, pp. 1490–1495, 2023

  60. [60]

    Cascaded state space and contrastive learning for cross-domain few-shot segmentation,

    F. Xiao, J. Zhang, P. Han, S. Chen, and H. Zhang, “Cascaded state space and contrastive learning for cross-domain few-shot segmentation,”IEEE Transactions on Industrial Informatics, 2025

  61. [61]

    Meta-learning based hyperspectral target detection using siamese network,

    Y. Wang, X. Chen, F. Wang, M. Song, and C. Yu, “Meta-learning based hyperspectral target detection using siamese network,”IEEE Trans. Geosci. Remote Sens., vol. 60, Art. no. 5527913, 2022

  62. [62]

    Background learning based on target suppression constraint for hyperspectral target detection,

    W. Xie, X. Zhang, Y. Li, K. Wang, and Q. Du, “Background learning based on target suppression constraint for hyperspectral target detection,”IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens., vol. 13, pp. 5887–5897, 2020

  63. [63]

    Hyperspectral target detection based on interpretable representation network,

    D. Shen, X. Ma, W. Kong, J. Liu, J. Wang, and H. Wang, “Hyperspectral target detection based on interpretable representation network,”IEEE Trans. Geosci. Remote Sens., vol. 61, Art. no. 5519416, 2023

  64. [64]

    Selfmtl: Self-supervised meta-transfer learning via contrastive representation for hyperspectral target detection,

    F. Luo, S. Shi, K. Qin, T. Guo, C. Fu, and Z. Lin, “Selfmtl: Self-supervised meta-transfer learning via contrastive representation for hyperspectral target detection,”IEEE Transactions on Geoscience and Remote Sensing, vol. 63, pp. 1–13, 2025

  65. [65]

    Htd-mamba: Efficient hyperspectral target detection with pyramid state space model,

    D. Shen, X. Zhu, J. Tian, J. Liu, Z. Du, H. Wang, and X. Ma, “Htd-mamba: Efficient hyperspectral target detection with pyramid state space model,” IEEE Transactions on Geoscience and Remote Sensing, vol. 63, pp. 1–15, 2025. 11 Appendix This supplementary material provides additional experimental evidence to complement the main paper. In particular, we fur...

  66. [66]

    These results provide a more complete understanding of the robustness and effectiveness of the proposed framework, and further support the conclusions drawn in the main text

    parameter sensitivity under the 10-way 2-shot protocol, including contextual, adaptation, and optimization-related hy- perparameters; and 2) ablation studies on the core modules and internal design choices. These results provide a more complete understanding of the robustness and effectiveness of the proposed framework, and further support the conclusions...