pith. machine review for the scientific record. sign in

arxiv: 2605.01691 · v1 · submitted 2026-05-03 · 💻 cs.LG

Recognition: 3 theorem links

· Lean Theorem

Complex Diffusion Maps with ω-Parameterized Kernels Revealing Inherent Harmonic Representations

Authors on Pith no claims yet

Pith reviewed 2026-05-08 19:30 UTC · model grok-4.3

classification 💻 cs.LG
keywords complex diffusion mapsomega parameterized kernelscomplex harmonicsdiffusion operatorangular structurespectral embeddingfMRI analysisEEG classification
0
0 comments X

The pith

Complex Diffusion Maps use ω-parameterized complex kernels to reveal dominant harmonic representations that preserve angular structure in high-dimensional data.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper proposes Complex Diffusion Maps to extract the main complex harmonics from data using a tunable family of complex-valued kernels. These kernels trade off local and nonlocal connections, drawing from both heat and Schrödinger equations to define a diffusion operator. The resulting maps are designed to keep angular information intact, which helps distinguish similar points better than real-valued approaches. This framework is shown to be robust in noisy environments and effective for analyzing brain activity signals.

Core claim

The central discovery is that a unified family of ω-parameterized complex-valued kernels defines diffusion operators whose spectra yield complex harmonic maps. These maps preserve angular structure in the complex plane rather than relying on magnitudes alone, and the complex kernel form amplifies distinctions among confusable samples. The approach establishes well-defined diffusion distances and provides an optimization interpretation for the embeddings.

What carries the argument

The ω-parameterized complex-valued kernel, which generates a diffusion operator and its complex eigenmaps that maintain angular relationships.

Load-bearing premise

The complex kernels produce diffusion operators with spectra that give meaningful complex harmonic maps preserving angular structure beyond real-valued kernels.

What would settle it

Observing no improvement in sample separation or eigengap clarity when applying the method to synthetic high-noise data with known confusable groups compared to standard diffusion maps would challenge the claim.

Figures

Figures reproduced from arXiv: 2605.01691 by Michael K. Ng, Tongzhen Dang, Weiyang Ding.

Figure 1
Figure 1. Figure 1: Motivation of the w-parameterized complex kernels and the interpretation of optimal embeddings. The Gaussian kernel in Diffusion Maps (DM) is related to the heat equation and mainly captures the local similarity of data. The complex-valued kernel induced by the Schr¨odinger equation has been shown to encode nonlocal connections through distance-dependent phase, allowing nonzero coupling and thus generating… view at source ↗
Figure 2
Figure 2. Figure 2: Illustration of the three-point toy example. (A) Euclidean distances (left), real-valued Gaussian kernel (middle), and angles between global connections (right). Points 1 (blue) and 2 (yellow) are close and indistinguishable under the real-valued kernel, while complex weights enhance their differences via angle information. (B) The first two embeddings from DM (left), the real parts of CDM (middle), and th… view at source ↗
Figure 3
Figure 3. Figure 3: Comparison of CDM with linear and nonlinear real-valued methods on an artificial dataset constructed via amplitude–phase coupling (see Appendix B for details). (A) Heatmap of the data matrix (left), corresponding Euclidean distances (middle), and angle distribution (right). Classes 1 and 3 exhibit similar distance structures and are difficult to distinguish, while Class 2 shows distinct intra-class distanc… view at source ↗
Figure 4
Figure 4. Figure 4: Low-dimensional representations of DM and CDM for a simulated time series under different noise levels, low noise ε = 0.1 (SNR ≈ 17 dB), medium noise ε = 0.5 (SNR ≈ 3 dB), and high noise ε = 1 (SNR ≈ −3 dB). Under low noise, the first two dimensions of DM capture only low-frequency components, whereas CDM reveals a clear four-cluster structure. As noise increases, the embeddings of both methods become more… view at source ↗
Figure 5
Figure 5. Figure 5: Eigenvalues and the first three dominant maps of DM and CDM for the simulated time series under high noise ε = 1 (SNR ≈ −3 dB). The eigenvalue gaps of CDM are more pronounced than those of DM, with the first three components contributing a larger proportion of the total variance. The corresponding three-dimensional embeddings reveal a clear four-cluster structure. In contrast, the real-valued kernel produc… view at source ↗
Figure 6
Figure 6. Figure 6: Comparison of functional connectivity (FC) reconstruction and edge-centric metastability (ECM) derived from resting-state fMRI signals of 100 participants from the Human Connectome Project [46] under PCA, DM, and CDM frameworks. (A) Correlation and reconstruction error between reconstructed and source FC matrices in the learned low-dimensional manifolds, evaluated over kernel bandwidth σ 2 and diffusion st… view at source ↗
Figure 7
Figure 7. Figure 7: Clustering results on the ISRUC-S3 dataset [49]. The dataset contains overnight polysomnography (PSG) recordings from 10 subjects, annotated into five sleep stages: Wake, N1, N2, N3, and rapid eye movement (REM). The evaluation metrics ACCclu, ARI, and NMI are described in D.3.1. (A) Mean clustering scores of CDM and DM under varying embedding dimensions and stacking orders p. CDM consistently achieves hig… view at source ↗
Figure 8
Figure 8. Figure 8: 10-fold cross-validation results of CDM under the intra-subject setting on the ISRUC-S3 dataset [49]. (A) Test accuracy ACCcls for one fold at stacking order p = 4, where dim denotes the embedding dimension. The best result is marked with a star. (B) Best test accuracy of each fold under different stacking orders p, with the mean accuracy indicated by hollow markers, showing the optimal order at p = 4. (C)… view at source ↗
Figure 9
Figure 9. Figure 9: Display of runtime on intra-subject setting. Runtime refers to the average runtime of 10-fold experiments with their respective optimal parameters. (A) Comparison of the runtime of CDM with baseline neural network-based methods. (B) Breakdown of CDM’s runtime across its four phases, Order-p data stacking, CDM computation, embedding alignment, and SVM classification. The results show that CDM outperforms se… view at source ↗
Figure 10
Figure 10. Figure 10: 10-fold cross-validation results of CDM under the cross-subject setting on the ISRUC-S3 dataset [49]. (A) Test accuracy ACCcls of one fold under the stacking order p = 15. The variable dim denotes the dimension of the low-dimensional embeddings used for SVM [72] classification. The best performance is marked with a star. (B) The best test accuracy of each fold under different order p, where colored scatte… view at source ↗
Figure 11
Figure 11. Figure 11: Display of runtime on cross-subject setting. Runtime refers to the average runtime of 10-fold experiments with their respective optimal parameters. (A) Comparison of the runtime of CDM with baselines. (B) Breakdown of CDM’s runtime across its four phases, Order-p data stacking, CDM computation, embedding alignment, and SVM classification. Combined with the lower part of view at source ↗
Figure 12
Figure 12. Figure 12: Display of runtime on intra-subject and cross-subject view at source ↗
read the original abstract

In this paper, we propose Complex Diffusion Maps (CDM), a novel diffusion mapping framework that aims to reveal the dominant complex harmonics of high-dimensional data. Inspired by the local Gaussian kernel relevant to the heat equation and the nonlocal Schr\"odinger kernel relevant to the Schr\"odinger equation, we propose a unified family of $\omega$-parameterized complex-valued kernels for the trade-off between local and nonlocal connections. We establish the theoretical foundation based on the operator spectrum theory, where the corresponding diffusion operator, diffusion distance, and complex harmonic maps are well-defined. An optimization-based interpretation of the maps is also developed, aiming to preserve angular structure in the complex diffusion space rather than relying solely on real-valued magnitude. We extensively evaluate CDM on both synthetic and real-world datasets. The complex-valued kernel amplifies differences among easily confusable samples, improving discriminative power over both linear and nonlinear methods based on real-valued kernels. CDM remains robust in high-noise settings, yielding a clearer eigengap that enhances spectral separation. For resting-state fMRI data, CDM captures more strongly correlated and nonlocal spatiotemporal dynamics. Without task-specific tuning, CDM achieves competitive performance on a public EEG sleep dataset, while maintaining high computational efficiency compared with both traditional machine learning and deep neural network approaches, highlighting its generality and practical value.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 2 minor

Summary. The paper proposes Complex Diffusion Maps (CDM) using a unified family of ω-parameterized complex-valued kernels that interpolate between local Gaussian kernels (heat equation) and nonlocal Schrödinger kernels. It claims to establish a theoretical foundation via operator spectrum theory under which the associated diffusion operator, diffusion distance, and complex harmonic maps are well-defined, with an optimization-based interpretation that preserves angular structure in the complex plane. Empirical evaluations on synthetic data, real-world datasets, resting-state fMRI, and a public EEG sleep dataset report improved discriminative power over real-valued linear and nonlinear methods, greater robustness in high-noise regimes with clearer eigengaps, and competitive performance with high computational efficiency.

Significance. If the extension of diffusion-map theory to complex kernels is rigorously justified and the reported empirical advantages hold under replication, the work could provide a practical tool for datasets exhibiting phase or angular structure, such as neuroimaging time series, while offering a parameter-controlled trade-off between local and nonlocal connectivity. The optimization view of angular preservation and the unified kernel family are potentially generalizable strengths.

major comments (2)
  1. [Abstract] Abstract: The claim that 'the corresponding diffusion operator, diffusion distance, and complex harmonic maps are well-defined' via operator spectrum theory is load-bearing for all downstream assertions (discriminative power, eigengap clarity, angular preservation). For a general complex kernel K(x,y) the induced integral operator need not be self-adjoint or normal on L², so the classical spectral theorem supplying a countable orthonormal basis of eigenfunctions with real eigenvalues does not apply directly; no explicit inner-product adjustment, Hermitian symmetrization, or compactness/normalcy proof is referenced to restore these properties.
  2. [Theoretical development (presumed §3–4)] Theoretical development (presumed §3–4): The optimization-based interpretation that the maps 'preserve angular structure in the complex diffusion space' presupposes that the complex spectrum yields geometrically meaningful harmonics beyond magnitude; without a concrete demonstration that the ω-parameterized kernel produces a compact normal operator (or equivalent) whose eigenfunctions retain angular interpretability, the superiority over real-valued kernels in confusable-sample and high-noise settings remains an unverified extension.
minor comments (2)
  1. [Abstract and Experiments] The abstract and evaluation sections would benefit from explicit statements of the precise values of ω used in each experiment and any sensitivity analysis, as the single free parameter is central to the claimed trade-off.
  2. [Experiments] Quantitative tables or figures reporting eigengap sizes, classification accuracies, or correlation strengths with error bars or statistical tests are needed to substantiate claims of 'clearer eigengap' and 'more strongly correlated' dynamics.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for their careful reading and constructive comments on our manuscript. We address each major comment below and commit to revisions that will strengthen the theoretical justifications as requested.

read point-by-point responses
  1. Referee: [Abstract] Abstract: The claim that 'the corresponding diffusion operator, diffusion distance, and complex harmonic maps are well-defined' via operator spectrum theory is load-bearing for all downstream assertions (discriminative power, eigengap clarity, angular preservation). For a general complex kernel K(x,y) the induced integral operator need not be self-adjoint or normal on L², so the classical spectral theorem supplying a countable orthonormal basis of eigenfunctions with real eigenvalues does not apply directly; no explicit inner-product adjustment, Hermitian symmetrization, or compactness/normalcy proof is referenced to restore these properties.

    Authors: We appreciate this observation. While our ω-parameterized kernels are constructed to satisfy the Hermitian symmetry condition K(y, x) = conjugate(K(x, y)), which ensures the integral operator is self-adjoint on the complex L² space, we acknowledge that the manuscript does not explicitly state the compactness and normality arguments. In the revised version, we will add a dedicated proposition in Section 3 (or an appendix) proving that, under standard assumptions on the data manifold being compact and the kernel being continuous and bounded, the operator is compact and self-adjoint, thereby justifying the application of the spectral theorem. This will include the explicit verification for the boundary cases (ω=0 corresponding to real Gaussian and ω=1 to Schrödinger-like) and the interpolation. We believe this addresses the concern without altering the core claims. revision: yes

  2. Referee: [Theoretical development (presumed §3–4)] Theoretical development (presumed §3–4): The optimization-based interpretation that the maps 'preserve angular structure in the complex diffusion space' presupposes that the complex spectrum yields geometrically meaningful harmonics beyond magnitude; without a concrete demonstration that the ω-parameterized kernel produces a compact normal operator (or equivalent) whose eigenfunctions retain angular interpretability, the superiority over real-valued kernels in confusable-sample and high-noise settings remains an unverified extension.

    Authors: We agree that a more explicit link between the operator properties and the angular interpretability would strengthen the paper. The optimization view is derived from viewing the embedding as minimizing a complex-valued loss that penalizes deviations in argument (phase) as well as magnitude. To demonstrate this, we will include in the revision a short subsection or example showing that for the proposed kernels, the eigenfunctions are complex-valued with phases that correspond to harmonic oscillations, and that the diffusion distance incorporates both real and imaginary parts. This will be supported by the spectral properties established in the new proposition. Regarding the empirical superiority, the experiments on synthetic data with confusable samples and high-noise regimes already illustrate the benefits, but we will add a note clarifying that these rely on the well-defined spectrum. We do not claim the superiority is solely due to the theory but is observed empirically when using the complex maps. revision: yes

Circularity Check

0 steps flagged

No significant circularity; derivation is self-contained.

full rationale

The paper defines a family of ω-parameterized complex kernels, invokes standard operator spectrum theory to define the diffusion operator/distance/maps, and provides an optimization view for angular preservation. All claims (eigengap, discriminative power, noise robustness) rest on these definitions plus empirical evaluation rather than reducing to fitted parameters renamed as predictions or to self-citation chains. No equation equates a derived quantity to its own input by construction, and no load-bearing uniqueness theorem is imported from the authors' prior work. The framework extends real-valued diffusion maps without internal circular reduction.

Axiom & Free-Parameter Ledger

1 free parameters · 1 axioms · 1 invented entities

The central claim depends on the existence of a well-defined spectrum for the complex diffusion operator and on the interpretability of angular preservation in the embedding space.

free parameters (1)
  • ω
    Tunable parameter balancing local Gaussian and nonlocal Schrödinger-type kernel contributions; its specific value selection is not detailed in the abstract.
axioms (1)
  • domain assumption Operator spectrum theory applies to the diffusion operator induced by the ω-parameterized complex kernel.
    Invoked to define diffusion distance and complex harmonic maps.
invented entities (1)
  • Complex harmonic maps no independent evidence
    purpose: Embeddings that preserve angular structure in complex diffusion space.
    New concept introduced to interpret the output of the complex operator beyond real magnitude.

pith-pipeline@v0.9.0 · 5536 in / 1257 out tokens · 58983 ms · 2026-05-08T19:30:47.433756+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Lean theorems connected to this paper

Citations machine-checked in the Pith Canon. Every link opens the source theorem in the public Lean library.

What do these tags mean?
matches
The paper's claim is directly supported by a theorem in the formal canon.
supports
The theorem supports part of the paper's argument, but the paper may add assumptions or extra steps.
extends
The paper goes beyond the formal theorem; the theorem is a base layer rather than the whole result.
uses
The paper appears to rely on the theorem as machinery.
contradicts
The paper's claim conflicts with a theorem or certificate in the canon.
unclear
Pith found a possible connection, but the passage is too broad, indirect, or ambiguous to say the theorem truly supports the claim.

Reference graph

Works this paper leans on

72 extracted references

  1. [1]

    Principal component analysis.Chemometrics and Intelligent Laboratory Systems, 2(1-3):37–52, 1987

    Svante Wold, Kim Esbensen, and Paul Geladi. Principal component analysis.Chemometrics and Intelligent Laboratory Systems, 2(1-3):37–52, 1987

  2. [2]

    Multidimensional scaling

    Mark L Davison and Stephen G Sireci. Multidimensional scaling. InHandbook of Applied Multivariate Statistics and Mathematical Modeling, pages 323–352. Elsevier, 2000

  3. [3]

    Kernel principal component analysis

    Bernhard Sch¨ olkopf, Alexander Smola, and Klaus-Robert M¨ uller. Kernel principal component analysis. InInternational Conference on Artificial Neural Networks, pages 583–588. Springer, 1997

  4. [4]

    A global geometric framework for nonlinear dimensionality reduction.Science, 290(5500):2319–2323, 2000

    Joshua B Tennenbaum, Vin de Silva, and John C Langford. A global geometric framework for nonlinear dimensionality reduction.Science, 290(5500):2319–2323, 2000

  5. [5]

    Nonlinear dimensionality reduction by locally linear embedding.Science, 290(5500):2323–2326, 2000

    Sam T Roweis and Lawrence K Saul. Nonlinear dimensionality reduction by locally linear embedding.Science, 290(5500):2323–2326, 2000

  6. [6]

    Laplacian eigenmaps for dimensionality reduction and data representation.Neural Computation, 15(6):1373–1396, 2003

    Mikhail Belkin and Partha Niyogi. Laplacian eigenmaps for dimensionality reduction and data representation.Neural Computation, 15(6):1373–1396, 2003

  7. [7]

    Ronald R Coifman, Stephane Lafon, Ann B Lee, Mauro Maggioni, Boaz Nadler, Frederick Warner, and Steven W Zucker. Geometric diffusions as a tool for harmonic analysis and structure definition of data: Diffusion maps.Proceedings of the National Academy of Sciences of the United States of America, 102(21):7426–7431, 2005

  8. [8]

    Visualizing data using t-sne.Journal of Machine Learning Research, 9(11), 2008

    Laurens Van der Maaten and Geoffrey Hinton. Visualizing data using t-sne.Journal of Machine Learning Research, 9(11), 2008

  9. [9]

    American Mathematical Society, 2022

    Lawrence C Evans.Partial differential equations. American Mathematical Society, 2022

  10. [10]

    Cambridge University Press, 1997

    Steven Rosenberg.The Laplacian on a Riemannian manifold: An introduction to analysis on manifolds. Cambridge University Press, 1997

  11. [11]

    The complex Gaussian kernel LMS algorithm

    Pantelis Bouboulis and Sergios Theodoridis. The complex Gaussian kernel LMS algorithm. In International Conference on Artificial Neural Networks, pages 11–20. Springer, 2010

  12. [12]

    Complex kernels for proper complex-valued signals: A review

    Rafael Boloix-Tortosa, F Javier Pay´ an-Somet, Eva Arias-de Reyna, and Juan Jos´ e Murillo- Fuentes. Complex kernels for proper complex-valued signals: A review. InEuropean Signal Processing Conference, pages 2371–2375. IEEE, 2015

  13. [13]

    Complex networks with complex weights.Physical Review E, 109(2):024314, 2024

    Lucas B¨ ottcher and Mason A Porter. Complex networks with complex weights.Physical Review E, 109(2):024314, 2024

  14. [14]

    Quantum theory based on real numbers can be experimentally falsified.Nature, 600(7890):625–629, 2021

    Marc-Olivier Renou, David Trillo, Mirjam Weilenmann, Thinh P Le, Armin Tavakoli, Nicolas Gisin, Antonio Ac´ ın, and Miguel Navascu´ es. Quantum theory based on real numbers can be experimentally falsified.Nature, 600(7890):625–629, 2021

  15. [15]

    Ruling out real-valued standard formalism of quantum theory.Physical Review Letters, 128(4):040403, 2022

    Ming-Cheng Chen, Can Wang, Feng-Ming Liu, Jian-Wen Wang, Chong Ying, Zhong-Xia Shang, Yulin Wu, Ming Gong, Hui Deng, F-T Liang, et al. Ruling out real-valued standard formalism of quantum theory.Physical Review Letters, 128(4):040403, 2022

  16. [16]

    Fluxes, Laplacians, and Kasteleyn’s theorem

    Elliott H Lieb and Michael Loss. Fluxes, Laplacians, and Kasteleyn’s theorem. InStatistical Mechanics: Selecta of Elliott H. Lieb, pages 457–483. Springer, 1993. 40

  17. [17]

    Magnetic eigenmaps for the visualization of directed networks.Applied and Computational Harmonic Analysis, 44(1):189–199, 2018

    Micha¨ el Fanuel, Carlos M Ala´ ız,´Angela Fern´ andez, and Johan AK Suykens. Magnetic eigenmaps for the visualization of directed networks.Applied and Computational Harmonic Analysis, 44(1):189–199, 2018

  18. [18]

    Magnet: A neural network for directed graphs

    Xitong Zhang, Yixuan He, Nathan Brugnone, Michael Perlmutter, and Matthew Hirn. Magnet: A neural network for directed graphs. InAdvances in Neural Information Processing Systems, pages 27003–27015, 2021

  19. [19]

    An optical neural chip for implementing complex-valued neural network.Nature Communications, 12(1):457, 2021

    Hui Zhang, Mile Gu, XD Jiang, Jayne Thompson, Hong Cai, Stefano Paesani, Raffaele Santagati, Anthony Laing, Y Zhang, Man-Hong Yung, et al. An optical neural chip for implementing complex-valued neural network.Nature Communications, 12(1):457, 2021

  20. [20]

    Complex-valued neural networks: A comprehensive survey.IEEE/CAA Journal of Automatica Sinica, 9(8):1406–1426, 2022

    ChiYan Lee, Hideyuki Hasegawa, and Shangce Gao. Complex-valued neural networks: A comprehensive survey.IEEE/CAA Journal of Automatica Sinica, 9(8):1406–1426, 2022

  21. [21]

    Springer, 1998

    Yuri A Kuznetsov, Iu A Kuznetsov, and Y Kuznetsov.Elements of applied bifurcation theory. Springer, 1998

  22. [22]

    Signal-coupled subthreshold Hopf-type systems show a sharpened collective response.Physical Review Letters, 116(10):108101, 2016

    Florian Gomez, Tom Lorimer, and Ruedi Stoop. Signal-coupled subthreshold Hopf-type systems show a sharpened collective response.Physical Review Letters, 116(10):108101, 2016

  23. [23]

    Complex harmonics reveal low-dimensional manifolds of critical brain dynamics.Physical Review E, 111(1):014410, 2025

    Gustavo Deco, Yonatan Sanz Perl, and Morten L Kringelbach. Complex harmonics reveal low-dimensional manifolds of critical brain dynamics.Physical Review E, 111(1):014410, 2025

  24. [24]

    Fractional Schr¨ odinger equation.Physical Review E, 66(5):056108, 2002

    Nick Laskin. Fractional Schr¨ odinger equation.Physical Review E, 66(5):056108, 2002

  25. [25]

    Fractional dynamics of systems with long-range interaction.Communications in Nonlinear Science and Numerical Simulation, 11(8):885–898, 2006

    Vasily E Tarasov and George M Zaslavsky. Fractional dynamics of systems with long-range interaction.Communications in Nonlinear Science and Numerical Simulation, 11(8):885–898, 2006

  26. [26]

    American Mathematical Society, 1990

    Walter A Strauss.Nonlinear wave equations. American Mathematical Society, 1990

  27. [27]

    Rare long-range cortical connections enhance human information processing.Current Biology, 31(20):4436–4448, 2021

    Gustavo Deco, Yonathan Sanz Perl, Peter Vuust, Enzo Tagliazucchi, Henry Kennedy, and Morten L Kringelbach. Rare long-range cortical connections enhance human information processing.Current Biology, 31(20):4436–4448, 2021

  28. [28]

    Non-local schr¨ odinger diffusion model reveals mechanisms of critical brain dynamics.Cell Reports Physical Science, 2025

    Gustavo Deco, Yonatan Sanz Perl, and Morten L Kringelbach. Non-local schr¨ odinger diffusion model reveals mechanisms of critical brain dynamics.Cell Reports Physical Science, 2025

  29. [29]

    Courier Corporation, 2013

    Naum Il’ich Akhiezer and Izrail Markovich Glazman.Theory of linear operators in Hilbert space. Courier Corporation, 2013

  30. [30]

    Vector diffusion maps and the connection Laplacian.Communications on Pure and Applied Mathematics, 65(8):1067–1144, 2012

    Amit Singer and H-T Wu. Vector diffusion maps and the connection Laplacian.Communications on Pure and Applied Mathematics, 65(8):1067–1144, 2012

  31. [31]

    Discrete magnetic Laplacian.Communications in Mathematical Physics, 164(2):259–275, 1994

    Mikhail A Shubin. Discrete magnetic Laplacian.Communications in Mathematical Physics, 164(2):259–275, 1994

  32. [32]

    Diffusion representation for asymmetric kernels.Applied Numerical Mathematics, 166:208–226, 2021

    Alvaro Almeida Gomez, Antˆ onio J Silva Neto, and Jorge P Zubelli. Diffusion representation for asymmetric kernels.Applied Numerical Mathematics, 166:208–226, 2021

  33. [33]

    Diffusion representation for asymmetric kernels via magnetic transform

    Mingzhen He, Fan He, Ruikai Yang, and Xiaolin Huang. Diffusion representation for asymmetric kernels via magnetic transform. InAdvances in Neural Information Processing Systems, 2023. 41

  34. [34]

    Gaussian filters for nonlinear filtering problems.IEEE Transac- tions on Automatic Control, 45(5):910–927, 2000

    Kazufumi Ito and Kaiqi Xiong. Gaussian filters for nonlinear filtering problems.IEEE Transac- tions on Automatic Control, 45(5):910–927, 2000

  35. [35]

    Asymptotic behaviors of support vector machines with Gaussian kernel.Neural Computation, 15(7):1667–1689, 2003

    S Sathiya Keerthi and Chih-Jen Lin. Asymptotic behaviors of support vector machines with Gaussian kernel.Neural Computation, 15(7):1667–1689, 2003

  36. [36]

    Convergence of Laplacian eigenmaps

    Mikhail Belkin and Partha Niyogi. Convergence of Laplacian eigenmaps. InAdvances in Neural Information Processing Systems, 2006

  37. [37]

    Towards a theoretical foundation for Laplacian-based manifold methods.Journal of Computer and System Sciences, 74(8):1289–1308, 2008

    Mikhail Belkin and Partha Niyogi. Towards a theoretical foundation for Laplacian-based manifold methods.Journal of Computer and System Sciences, 74(8):1289–1308, 2008

  38. [38]

    Long range arena: A benchmark for efficient transformers

    Yi Tay, Mostafa Dehghani, Samira Abnar, Yikang Shen, Dara Bahri, Philip Pham, Jinfeng Rao, Liu Yang, Sebastian Ruder, and Donald Metzler. Long range arena: A benchmark for efficient transformers. InInternational Conference on Learning Representations, 2021

  39. [39]

    Cambridge University Press, 2018

    David J Griffiths and Darrell F Schroeter.Introduction to quantum mechanics. Cambridge University Press, 2018

  40. [40]

    Princeton University Press, 2010

    Anthony Zee.Quantum field theory in a nutshell. Princeton University Press, 2010

  41. [41]

    Complex symmetric operators and applications.Transactions of the American Mathematical Society, 358(3):1285–1315, 2006

    Stephan Garcia and Mihai Putinar. Complex symmetric operators and applications.Transactions of the American Mathematical Society, 358(3):1285–1315, 2006

  42. [42]

    Complex symmetric operators and applications ii.Transac- tions of the American Mathematical Society, 359(8):3913–3931, 2007

    Stephan Garcia and Mihai Putinar. Complex symmetric operators and applications ii.Transac- tions of the American Mathematical Society, 359(8):3913–3931, 2007

  43. [43]

    Springer Science & Business Media, 1994

    John B Conway.A course in functional analysis. Springer Science & Business Media, 1994

  44. [44]

    American Mathematical Society, 1997

    Fan RK Chung.Spectral graph theory. American Mathematical Society, 1997

  45. [45]

    Using the Nystr¨ om method to speed up kernel machines

    Christopher Williams and Matthias Seeger. Using the Nystr¨ om method to speed up kernel machines. InAdvances in Neural Information Processing Systems, 2000

  46. [46]

    The wu-minn Human Connectome Project: An overview.Neuroimage, 80:62–79, 2013

    David C Van Essen, Stephen M Smith, Deanna M Barch, Timothy EJ Behrens, Essa Yacoub, Kamil Ugurbil, Wu-Minn HCP Consortium, et al. The wu-minn Human Connectome Project: An overview.Neuroimage, 80:62–79, 2013

  47. [47]

    A multi-modal parcellation of human cerebral cortex.Nature, 536(7615):171–178, 2016

    Matthew F Glasser, Timothy S Coalson, Emma C Robinson, Carl D Hacker, John Harwell, Essa Yacoub, Kamil Ugurbil, Jesper Andersson, Christian F Beckmann, Mark Jenkinson, et al. A multi-modal parcellation of human cerebral cortex.Nature, 536(7615):171–178, 2016

  48. [48]

    Edge-centric functional network representations of human cerebral cortex reveal overlapping system-level architecture.Nature Neuroscience, 23(12):1644–1654, 2020

    Joshua Faskowitz, Farnaz Zamani Esfahlani, Youngheun Jo, Olaf Sporns, and Richard F Betzel. Edge-centric functional network representations of human cerebral cortex reveal overlapping system-level architecture.Nature Neuroscience, 23(12):1644–1654, 2020

  49. [49]

    ISRUC-Sleep: A comprehensive public dataset for sleep researchers.Computer Methods and Programs in Biomedicine, 124:180–192, 2016

    Sirvan Khalighi, Teresa Sousa, Jos´ e Moutinho Santos, and Urbano Nunes. ISRUC-Sleep: A comprehensive public dataset for sleep researchers.Computer Methods and Programs in Biomedicine, 124:180–192, 2016

  50. [50]

    Geometry from a time series.Physical Review Letters, 45(9):712, 1980

    Norman H Packard, James P Crutchfield, J Doyne Farmer, and Robert S Shaw. Geometry from a time series.Physical Review Letters, 45(9):712, 1980. 42

  51. [51]

    Detecting strange attractors in turbulence

    Floris Takens. Detecting strange attractors in turbulence. InDynamical Systems and Turbulence, Warwick 1980, pages 366–381. Springer, 2006

  52. [52]

    Unsupervised learning

    Trevor Hastie, Robert Tibshirani, and Jerome Friedman. Unsupervised learning. InThe Elements of Statistical Learning: Data Mining, Inference, and Prediction, pages 485–585. Springer, 2008

  53. [53]

    Supervised learning

    P´ adraig Cunningham, Matthieu Cord, and Sarah Jane Delany. Supervised learning. InMachine Learning Techniques for Multimedia: Case Studies on Organization and Retrieval, pages 21–49. Springer, 2008

  54. [54]

    Multivariate observations

    J MacQueen. Multivariate observations. InBerkeley Symposium on Mathematical Statisticsand Probability, pages 281–297. University of California Press Oakland, 1967

  55. [55]

    Neural manifolds for the control of movement.Neuron, 94(5):978–984, 2017

    Juan A Gallego, Matthew G Perich, Lee E Miller, and Sara A Solla. Neural manifolds for the control of movement.Neuron, 94(5):978–984, 2017

  56. [56]

    A neural manifold view of the brain

    Matthew G Perich, Devika Narain, and Juan A Gallego. A neural manifold view of the brain. Nature Neuroscience, 28(8):1582–1597, 2025

  57. [57]

    SalientSleepNet: Multimodal salient wave detection network for sleep staging

    Ziyu Jia, Youfang Lin, Jing Wang, Xuehui Wang, Peiyi Xie, and Yingbin Zhang. SalientSleepNet: Multimodal salient wave detection network for sleep staging. InInternational Joint Conference on Artificial Intelligence, pages 2614–2620, 8 2021

  58. [58]

    Ziyu Jia, Youfang Lin, Jing Wang, Xiaojun Ning, Yuanlai He, Ronghao Zhou, Yuhan Zhou, and Li-wei H Lehman. Multi-view spatial-temporal graph convolutional networks with domain gener- alization for sleep stage classification.IEEE Transactions on Neural Systems and Rehabilitation Engineering, 29:1977–1986, 2021

  59. [59]

    Emadeldeen Eldele, Zhenghua Chen, Chengyu Liu, Min Wu, Chee-Keong Kwoh, Xiaoli Li, and Cuntai Guan. An attention-based deep learning approach for sleep stage classification with single-channel EEG.IEEE Transactions on Neural Systems and Rehabilitation Engineering, 29:809–818, 2021

  60. [60]

    Ziyu Jia, Xiyang Cai, Gaoxing Zheng, Jing Wang, and Youfang Lin. SleepPrintnet: A multivariate multimodal neural network based on physiological time-series for automatic sleep staging.IEEE Transactions on Artificial Intelligence, 1(3):248–257, 2021

  61. [61]

    Kipf and Max Welling

    Thomas N. Kipf and Max Welling. Semi-supervised classification with graph convolutional networks. InInternational Conference on Learning Representations, 2017

  62. [62]

    The minimal preprocessing pipelines for the Human Connectome Project.Neuroimage, 80:105–124, 2013

    Matthew F Glasser, Stamatios N Sotiropoulos, J Anthony Wilson, Timothy S Coalson, Bruce Fischl, Jesper L Andersson, Junqian Xu, Saad Jbabdi, Matthew Webster, Jonathan R Polimeni, et al. The minimal preprocessing pipelines for the Human Connectome Project.Neuroimage, 80:105–124, 2013

  63. [63]

    Geometric harmonics: A novel tool for multiscale out-of-sample extension of empirical functions.Applied and Computational Harmonic Analysis, 21(1):31–52, 2006

    Ronald R Coifman and St´ ephane Lafon. Geometric harmonics: A novel tool for multiscale out-of-sample extension of empirical functions.Applied and Computational Harmonic Analysis, 21(1):31–52, 2006

  64. [64]

    Time-series forecasting using manifold learning, radial basis function interpolation, and geometric harmonics.Chaos: An Interdisciplinary Journal of Nonlinear Science, 32(8), 2022

    Panagiotis G Papaioannou, Ronen Talmon, Ioannis G Kevrekidis, and Constantinos Siettos. Time-series forecasting using manifold learning, radial basis function interpolation, and geometric harmonics.Chaos: An Interdisciplinary Journal of Nonlinear Science, 32(8), 2022. 43

  65. [65]

    Data-driven control of agent-based models: An equation/variable-free machine learning approach.Journal of Computational Physics, 478:111953, 2023

    Dimitrios G Patsatzis, Lucia Russo, Ioannis G Kevrekidis, and Constantinos Siettos. Data-driven control of agent-based models: An equation/variable-free machine learning approach.Journal of Computational Physics, 478:111953, 2023

  66. [66]

    A novel multi-class EEG-based sleep stage classification system.IEEE Transactions on Neural Systems and Rehabilitation Engineering, 26(1):84–95, 2017

    Pejman Memar and Farhad Faradji. A novel multi-class EEG-based sleep stage classification system.IEEE Transactions on Neural Systems and Rehabilitation Engineering, 26(1):84–95, 2017

  67. [67]

    Ensemble svm method for automatic sleep stage classification.IEEE Transactions on Instrumentation and Measurement, 67(6):1258–1265, 2018

    Emina Alickovic and Abdulhamit Subasi. Ensemble svm method for automatic sleep stage classification.IEEE Transactions on Instrumentation and Measurement, 67(6):1258–1265, 2018

  68. [68]

    Mixed neural network approach for temporal sleep stage classification.IEEE Transactions on Neural Systems and Rehabilitation Engineering, 26(2):324–333, 2017

    Hao Dong, Akara Supratak, Wei Pan, Chao Wu, Paul M Matthews, and Yike Guo. Mixed neural network approach for temporal sleep stage classification.IEEE Transactions on Neural Systems and Rehabilitation Engineering, 26(2):324–333, 2017

  69. [69]

    Comparing partitions.Journal of Classification, 2(1):193– 218, 1985

    Lawrence Hubert and Phipps Arabie. Comparing partitions.Journal of Classification, 2(1):193– 218, 1985

  70. [70]

    Information theoretic measures for clusterings comparison: Is a correction for chance necessary? InInternational Conference on Machine Learning, pages 1073–1080, 2009

    Nguyen Xuan Vinh, Julien Epps, and James Bailey. Information theoretic measures for clusterings comparison: Is a correction for chance necessary? InInternational Conference on Machine Learning, pages 1073–1080, 2009

  71. [71]

    The use of multiple measurements in taxonomic problems.Annals of Eugenics, 7(2):179–188, 1936

    Ronald A Fisher. The use of multiple measurements in taxonomic problems.Annals of Eugenics, 7(2):179–188, 1936

  72. [72]

    Support-vector networks.Machine Learning, 20:273–297, 1995

    Corinna Cortes and Vladimir Vapnik. Support-vector networks.Machine Learning, 20:273–297, 1995. 44