pith. machine review for the scientific record. sign in

arxiv: 2604.25631 · v1 · submitted 2026-04-28 · 🪐 quant-ph

Recognition: unknown

Local tensor-train surrogates for quantum learning models

Christopher Ferrie, Sreeraj Rajindran Nair

Authors on Pith no claims yet

Pith reviewed 2026-05-07 16:54 UTC · model grok-4.3

classification 🪐 quant-ph
keywords quantum machine learningtensor trainsurrogate modelTaylor approximationempirical risk minimizationgeneralization boundslocal patchesquantum circuits
0
0 comments X

The pith

Local patches of quantum machine learning models admit fast classical tensor-train surrogates with explicit error bounds.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper develops a method to create classical surrogates for quantum machine learning models that are accurate only within small local regions of the input space. It uses Taylor polynomial expansions to approximate the model smoothly and then represents those polynomials efficiently with tensor trains. Empirical risk minimization is applied to learn the surrogate parameters from data, yielding provable generalization guarantees. This matters because quantum models are expensive to evaluate repeatedly, and the local surrogates offer a cheaper alternative while controlling three separate sources of error through patch size, polynomial degree, bond dimension, and sample count. The parameter scaling stays polynomial in the input dimension instead of exponential.

Core claim

The Taylor-TT construction serves as a deterministic error certificate proving that the TT hypothesis class contains a good approximation to the quantum model inside a local patch; empirical risk minimization then provably recovers a surrogate with controlled generalization error and explicit bounds on the risk. This yields three independently controllable error sources from Taylor truncation controlled by patch radius r and degree p, TT approximation controlled by bond dimension chi, and statistical estimation. Parameter count scales as N(p+1)chi squared rather than exponentially in N, cleanly separating representation complexity from the exponential constants induced by the tensor-product

What carries the argument

The Taylor-TT construction, which embeds a local Taylor polynomial approximation of the quantum model into a low-bond-dimension tensor-train representation for use as a hypothesis class in empirical risk minimization.

Load-bearing premise

The quantum model function is sufficiently smooth inside each local patch for a low-degree Taylor polynomial plus low-bond-dimension tensor-train to achieve useful accuracy, with the feature map permitting manageable tensor-product norms.

What would settle it

Collecting data inside a chosen local patch and training the TT surrogate; if the observed test error inside the patch does not decrease when the polynomial degree p or bond dimension chi is increased while keeping other parameters fixed, the error control claims would be falsified.

Figures

Figures reproduced from arXiv: 2604.25631 by Christopher Ferrie, Sreeraj Rajindran Nair.

Figure 1
Figure 1. Figure 1: Sample illustration of the local tensor-train surrogates (LTTS): Starting from a function view at source ↗
Figure 2
Figure 2. Figure 2: Scatter of minimal TT rank caps (χbox, χ∆) for 176 test cases across four (N, p) configurations. Top row: common-scale criterion ∥A − ATT∥F ≤ ε∥Abox∥F ; bottom row: self-relative criterion ∥A − ATT∥F ≤ ε∥A∥F . Left column: ε = 10−2 ; right column: ε = 10−3 . Points above the dashed diagonal indicate rank inflation (ρ > 1); points below indicate deflation (ρ < 1). Small random jitter is added for visibility… view at source ↗
Figure 3
Figure 3. Figure 3: Local TT surrogate (LTTS) validation across two datasets: Top row: synthetic Gaussian classification. Bottom row: UCI Banknote Authentication. Each row shows the same 4-panel diagnostic: (A) error decomposition vs patch radius r at the primary center x0, including an r 3 reference, (B) TT coefficient-tensor truncation error ∥C − Cχ∥F /∥C∥F vs rank cap χ across multiple centers; (C) total surrogate RMSE ∥g … view at source ↗
read the original abstract

A key bottleneck in quantum machine learning is the computational cost of repeated quantum circuit evaluations during the inference phase. To address this, we present a framework for constructing fast, cheap, provably accurate classical tensor-train surrogates of fully trained quantum machine learning models within local patches of their input data space. The approach combines Taylor polynomial approximation with a tensor-train (TT) representation and embeds it in a statistical learning paradigm via empirical risk minimization. In our analysis, the Taylor-TT construction serves as a deterministic error certificate proving that the TT hypothesis class contains a good approximation; empirical risk minimization then provably recovers a surrogate with controlled generalization error and explicit bounds. This translates into three independently controllable error sources: (i) Taylor truncation error controlled by the patch radius $r$ and polynomial degree $p$, (ii) TT approximation error controlled by the bond dimension $\chi$, and (iii) statistical estimation error. While the parameter count scales polynomially in the number of data dimensions $N$, i.e., $d_{\mathrm{eff}} = N(p+1)\chi^2$ rather than the naive $(p+1)^N$, the worst-case constants inherit an exponential factor through the tensor-product feature norm during Taylor polynomial embedding onto TT. This cleanly separates representation complexity from feature-induced constants. Our risk bounds and sample complexity depend explicitly on the local patch radius $r$.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 2 minor

Summary. The manuscript proposes constructing fast classical tensor-train (TT) surrogates for fully trained quantum machine learning models inside local patches of the input domain. It combines a Taylor polynomial expansion (controlled by patch radius r and degree p) with a low-bond-dimension TT representation (controlled by bond dimension χ) and embeds the construction in an empirical risk minimization (ERM) framework. The central claim is that the Taylor-TT approximant supplies a deterministic certificate that the TT hypothesis class contains a sufficiently accurate surrogate; standard uniform-convergence arguments then yield explicit generalization bounds whose three error terms (Taylor truncation, TT truncation, and statistical) are independently controllable. The effective parameter count scales as d_eff = N(p+1)χ² rather than exponentially in N, while the worst-case constants inherit an exponential factor from the tensor-product feature norm.

Significance. If the derivations and assumptions hold, the work supplies a concrete route to replace repeated quantum-circuit evaluations with cheap classical surrogates that come with explicit, tunable error certificates. The clean separation of representation complexity from feature-induced constants and the explicit dependence of sample complexity on the local radius r are useful features for practical QML deployment. The reliance on standard results from approximation theory and statistical learning theory keeps the argument modular and falsifiable.

major comments (2)
  1. [§3.2] §3.2, the statement that the TT approximation error is controlled solely by χ: the proof sketch must explicitly bound the TT truncation error for the multivariate Taylor polynomial under the tensor-product norm; without this quantitative link the claim that the three error sources are independently controllable is not yet load-bearing.
  2. [Theorem 4.1] Theorem 4.1 (generalization bound): the exponential factor arising from the tensor-product feature norm appears in the final sample-complexity expression; the manuscript should state whether this factor can be absorbed into the choice of r or whether it remains an unavoidable dependence on the feature map.
minor comments (2)
  1. [§2.1] The notation for the local patch radius r is introduced in the abstract but first defined only in §2.1; a forward reference or early definition would improve readability.
  2. [Figure 1] Figure 1 caption should clarify whether the plotted error curves are theoretical bounds or numerical realizations of the Taylor-TT surrogate.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the careful reading of our manuscript and the constructive comments. We appreciate the positive assessment of the overall approach and have revised the paper to address the points raised, strengthening the rigor of the error analysis and clarifying the dependence on the feature map.

read point-by-point responses
  1. Referee: [§3.2] §3.2, the statement that the TT approximation error is controlled solely by χ: the proof sketch must explicitly bound the TT truncation error for the multivariate Taylor polynomial under the tensor-product norm; without this quantitative link the claim that the three error sources are independently controllable is not yet load-bearing.

    Authors: We agree that an explicit quantitative link is required to fully substantiate the independent controllability of the three error terms. In the revised §3.2 we have augmented the proof sketch with a direct bound on the TT truncation error for the multivariate Taylor polynomial. The Taylor polynomial is first embedded into the tensor-product feature space; each monomial term admits an exact rank-1 TT representation, after which standard TT truncation results (e.g., via successive SVDs) yield an error controlled by χ, the polynomial degree p, and the tensor-product norm of the feature map. This bound depends only on representation parameters (χ, p) and the fixed feature norm, remaining independent of the statistical sample size. The revised text now states the three error sources explicitly and shows how each can be tuned separately. revision: yes

  2. Referee: [Theorem 4.1] Theorem 4.1 (generalization bound): the exponential factor arising from the tensor-product feature norm appears in the final sample-complexity expression; the manuscript should state whether this factor can be absorbed into the choice of r or whether it remains an unavoidable dependence on the feature map.

    Authors: The exponential factor arises from the tensor-product structure of the feature map norm evaluated inside the local patch; it scales as O(C^N) for a constant C determined by the feature map and the ambient input domain. While a smaller patch radius r reduces the Taylor truncation error (by suppressing higher-order remainder terms), it does not cancel or absorb the feature-norm prefactor, which is independent of r and stems directly from the product structure of the embedding. Consequently the factor remains an unavoidable dependence on the choice of feature map. We have added a clarifying remark immediately after the statement of Theorem 4.1 that makes this dependence explicit and notes that, for many physically motivated feature maps, the constant C can be moderate or further controlled by normalization. revision: yes

Circularity Check

0 steps flagged

No significant circularity; derivation is self-contained against external benchmarks

full rationale

The paper's central argument constructs a Taylor-TT surrogate as an existence certificate drawn from classical multivariate approximation theory (Taylor expansion with remainder controlled by patch radius r and degree p) and then invokes standard uniform convergence results from statistical learning theory to bound the generalization error of ERM over the TT hypothesis class. The effective dimension d_eff = N(p+1)χ² is derived directly from the TT parameterization without fitting to the target data, and the three error sources (truncation, TT compression, statistical) are controlled by independent parameters whose bounds do not reduce to any quantity defined by the same empirical risk minimization step. No self-citation is load-bearing for the existence or bound statements, and the tensor-product norm exponential factor is acknowledged as a worst-case constant inherited from the feature map rather than a fitted or redefined quantity. The derivation therefore remains independent of the quantum model specifics beyond the smoothness assumption inside each local patch.

Axiom & Free-Parameter Ledger

0 free parameters · 2 axioms · 0 invented entities

The framework rests on standard results from multivariate Taylor approximation and statistical learning theory; no new free parameters are fitted inside the central claim and no new physical entities are postulated.

axioms (2)
  • domain assumption Taylor's theorem applies to the quantum model function inside each local patch of radius r
    Invoked to bound truncation error by choice of polynomial degree p and radius r.
  • domain assumption A tensor-train decomposition of controlled bond dimension can approximate the embedded Taylor polynomial to arbitrary accuracy
    Used to separate TT approximation error from the other two sources.

pith-pipeline@v0.9.0 · 5544 in / 1387 out tokens · 35784 ms · 2026-05-07T16:54:14.427127+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

66 extracted references · 54 canonical work pages · 1 internal anchor

  1. [1]

    Su Yeon Chang and M. Cerezo. A primer on quantum machine learning, 2025. URL:https://arxiv. org/abs/2511.15969,arXiv:2511.15969

  2. [2]

    Pradeep Lamichhane and Danda B. Rawat. Quantum machine learning: Recent advances, challenges, and perspectives.IEEE Access, 13:94057–94105, 2025.doi:10.1109/ACCESS.2025.3573244

  3. [3]

    Andrew D

    Giovanni Acampora, Andris Ambainis, Natalia Ares, Leonardo Banchi, Pallavi Bhardwaj, Daniele Binosi, G. Andrew D. Briggs, Tommaso Calarco, Vedran Dunjko, Jens Eisert, Olivier Ezratty, Paul Erker, Federico Fedele, Elies Gil-Fuster, Martin G¨ arttner, Mats Granath, Markus Heyl, Iordanis Kerenidis, Matthias Klusch, Anton Frisk Kockum, Richard Kueng, Mario Kr...

  4. [4]

    A survey of quantum learning theory, 2017

    Srinivasan Arunachalam and Ronald de Wolf. A survey of quantum learning theory, 2017. URL: https://arxiv.org/abs/1701.06806,arXiv:1701.06806

  5. [5]

    The theory of variational hybrid quantum-classical algorithms.New Journal of Physics, 18(2):023023, February

    Jarrod R McClean, Jonathan Romero, Ryan Babbush, and Al´ an Aspuru-Guzik. The theory of variational hybrid quantum-classical algorithms.New Journal of Physics, 18(2):023023, February

  6. [6]

    URL:http://dx.doi.org/10.1088/1367-2630/18/2/023023,doi:10.1088/1367-2630/18/ 2/023023

  7. [7]

    Mitarai, M

    K. Mitarai, M. Negoro, M. Kitagawa, and K. Fujii. Quantum circuit learning.Phys. Rev. A, 98:032309, Sep 2018. URL:https://link.aps.org/doi/10.1103/PhysRevA.98.032309,doi: 10.1103/PhysRevA.98.032309

  8. [8]

    Vedran Dunjko and Hans J. Briegel. Machine learning & artificial intelligence in the quantum domain,

  9. [9]

    URL:https://arxiv.org/abs/1709.02779,arXiv:1709.02779

  10. [10]

    A comprehensive review of quantum machine learning: from nisq to fault tolerance.Reports on Progress in Physics, 87(11):116402, October 2024

    Yunfei Wang and Junyu Liu. A comprehensive review of quantum machine learning: from nisq to fault tolerance.Reports on Progress in Physics, 87(11):116402, October 2024. URL:http://dx.doi. org/10.1088/1361-6633/ad7f69,doi:10.1088/1361-6633/ad7f69

  11. [11]

    Quantum computing in the NISQ era and beyond.Quantum, 2:79, 2018

    John Preskill. Quantum computing in the nisq era and beyond.Quantum, 2:79, August 2018. URL: http://dx.doi.org/10.22331/q-2018-08-06-79,doi:10.22331/q-2018-08-06-79

  12. [12]

    Kottmann, Tim Menke, Wai-Keong Mok, Sukin Sim, Leong-Chuan Kwek, and Al´ an Aspuru-Guzik

    Kishor Bharti, Alba Cervera-Lierta, Thi Ha Kyaw, Tobias Haug, Sumner Alperin-Lea, Abhinav Anand, Matthias Degroote, Hermanni Heimonen, Jakob S. Kottmann, Tim Menke, Wai-Keong Mok, Sukin Sim, Leong-Chuan Kwek, and Al´ an Aspuru-Guzik. Noisy intermediate-scale quantum 21 algorithms.Rev. Mod. Phys., 94:015004, Feb 2022. URL:https://link.aps.org/doi/10.1103/ ...

  13. [13]

    Quantum machine learning on near-term quantum devices: Current state of supervised and unsupervised techniques for real-world applications,

    Yaswitha Gujju, Atsushi Matsuo, and Rudy Raymond. Quantum machine learning on near-term quantum devices: Current state of supervised and unsupervised techniques for real-world applications,

  14. [14]

    URL:https://arxiv.org/abs/2307.00908,arXiv:2307.00908

  15. [15]

    Quantum machine learning: from physics to software engineering.Advances in Physics: X, 8(1), February

    Alexey Melnikov, Mohammad Kordzanganeh, Alexander Alodjants, and Ray-Kuang Lee. Quantum machine learning: from physics to software engineering.Advances in Physics: X, 8(1), February

  16. [16]

    URL:http://dx.doi.org/10.1080/23746149.2023.2165452,doi:10.1080/23746149.2023. 2165452

  17. [17]

    Larocca, S

    Mart´ ın Larocca, Supanut Thanasilp, Samson Wang, Kunal Sharma, Jacob Biamonte, Patrick J. Coles, Lukasz Cincio, Jarrod R. McClean, Zo¨ e Holmes, and M. Cerezo. Barren plateaus in variational quantum computing.Nature Reviews Physics, 7(4):174–189, March 2025. URL:http://dx.doi. org/10.1038/s42254-025-00813-9,doi:10.1038/s42254-025-00813-9

  18. [18]

    Cerezo, Martin Larocca, Diego Garc´ ıa-Mart´ ın, N

    M. Cerezo, Martin Larocca, Diego Garc´ ıa-Mart´ ın, N. L. Diaz, Paolo Braccia, Enrico Fontana, Manuel S. Rudolph, Pablo Bermejo, Aroosa Ijaz, Supanut Thanasilp, Eric R. Anschuetz, and Zo¨ e Holmes. Does provable absence of barren plateaus imply classical simulability? or, why we need to rethink variational quantum computing, 2024. URL:https://arxiv.org/ab...

  19. [19]

    Mind the gaps: The fraught road to quantum advantage

    Jens Eisert and John Preskill. Mind the gaps: The fraught road to quantum advantage, 2025. URL: https://arxiv.org/abs/2510.19928,arXiv:2510.19928

  20. [20]

    ACM Transactions on Quantum Computing , year=

    John Preskill. Beyond nisq: The megaquop machine, 2025. URL:https://arxiv.org/abs/2502. 17368,arXiv:2502.17368,doi:10.1145/3723153

  21. [21]

    Zolt´ an Zimbor´ as, B´ alint Koczor, Zo¨ e Holmes, Elsi-Mari Borrelli, Andr´ as Gily´ en, Hsin-Yuan Huang, Zhenyu Cai, Antonio Ac´ ın, Leandro Aolita, Leonardo Banchi, Fernando G. S. L. Brand˜ ao, Daniel Cavalcanti, Toby Cubitt, Sergey N. Filippov, Guillermo Garc´ ıa-P´ erez, John Goold, Orsolya K´ alm´ an, Elica Kyoseva, Matteo A. C. Rossi, Boris Sokolo...

  22. [22]

    Generalization in quantum machine learning from few training data

    Manuel Caro, Hsin-Yuan Huang, M. Cerezo, Kunal Sharma, Andrew Sornborger, Lukasz Cincio, and Patrick J. Coles. Generalization in quantum machine learning from few training data.Nature Communications, 13:4919, 2022.doi:10.1038/s41467-022-32550-3

  23. [23]

    Marshall, and Vedran Dunjko

    Riccardo Molteni, Simon C. Marshall, and Vedran Dunjko. Quantum machine learning advantages be- yond hardness of evaluation, 2025. URL:https://arxiv.org/abs/2504.15964,arXiv:2504.15964

  24. [24]

    Carbon efficient quantum AI: an empirical study of ans¨ atz design trade-offs in QNN and QLSTM models.Sci

    Sarvapriya Tripathi, Himanshu Upadhyay, and Jayesh Soni. Carbon efficient quantum AI: an empirical study of ans¨ atz design trade-offs in QNN and QLSTM models.Sci. Rep., 15(1):44936, 2025.doi: 10.1038/s41598-025-28582-6

  25. [25]

    The resource cost of large scale quantum computing, 2022

    Marco Fellous-Asiani. The resource cost of large scale quantum computing, 2022. URL:https: //arxiv.org/abs/2112.04022,arXiv:2112.04022

  26. [26]

    Has quantum advantage been achieved?, 2026

    Dominik Hangleiter. Has quantum advantage been achieved?, 2026. URL:https://arxiv.org/abs/ 2603.09901,arXiv:2603.09901

  27. [27]

    Provable advantage in quantum pac learning, 2023

    Wilfred Salmon, Sergii Strelchuk, and Tom Gur. Provable advantage in quantum pac learning, 2023. URL:https://arxiv.org/abs/2309.10887,arXiv:2309.10887. 22

  28. [28]

    McClean, and John Preskill

    Hsin-Yuan Huang, Soonwon Choi, Jarrod R. McClean, and John Preskill. The vast world of quantum advantage, 2025. URL:https://arxiv.org/abs/2508.05720,arXiv:2508.05720

  29. [29]

    Is quantum advantage the right goal for quantum machine learn- ing?PRX Quantum, 3(3), July 2022

    Maria Schuld and Nathan Killoran. Is quantum advantage the right goal for quantum machine learn- ing?PRX Quantum, 3(3), July 2022. URL:http://dx.doi.org/10.1103/PRXQuantum.3.030101, doi:10.1103/prxquantum.3.030101

  30. [30]

    Jordan Cotler, Hsin-Yuan Huang, and Jarrod R. McClean. Revisiting dequantization and quantum advantage in learning tasks, 2021. URL:https://arxiv.org/abs/2112.00811,arXiv:2112.00811

  31. [31]

    Adri´ an P´ erez-Salinas, Alba Cervera-Lierta, Elies Gil-Fuster, and Jos´ e I. Latorre. Data re-uploading for a universal quantum classifier.Quantum, 4:226, February 2020. URL:http://dx.doi.org/10. 22331/q-2020-02-06-226,doi:10.22331/q-2020-02-06-226

  32. [32]

    Forn-D´ ıaz, and Jos´ e I

    Adri´ an P´ erez-Salinas, David L´ opez-N´ u˜ nez, Artur Garc´ ıa-S´ aez, P. Forn-D´ ıaz, and Jos´ e I. Latorre. One qubit as a universal approximant.Physical Review A, 104(1), July 2021. URL:http://dx.doi.org/ 10.1103/PhysRevA.104.012405,doi:10.1103/physreva.104.012405

  33. [33]

    doi:10.1103/physreva.103.032430

    Maria Schuld, Ryan Sweke, and Johannes Jakob Meyer. Effect of data encoding on the expressive power of variational quantum-machine-learning models.Phys. Rev. A, 103:032430, Mar 2021. URL:https: //link.aps.org/doi/10.1103/PhysRevA.103.032430,doi:10.1103/PhysRevA.103.032430

  34. [34]

    Multidimensional fourier series with quantum circuits.Phys

    Berta Casas and Alba Cervera-Lierta. Multidimensional fourier series with quantum circuits.Phys. Rev. A, 107:062612, Jun 2023. URL:https://link.aps.org/doi/10.1103/PhysRevA.107.062612, doi:10.1103/PhysRevA.107.062612

  35. [35]

    Schreiber, Jens Eisert, and Johannes Jakob Meyer

    Franz J. Schreiber, Jens Eisert, and Johannes Jakob Meyer. Classical surrogates for quantum learning models.Physical Review Letters, 131(10), September 2023. URL:http://dx.doi.org/10.1103/ PhysRevLett.131.100803,doi:10.1103/physrevlett.131.100803

  36. [36]

    Enhancing the scalability of classical surrogates for real-world quantum machine learning applications, 2025

    Philip Anton Hernicht, Alona Sakhnenko, Corey O’Meara, Giorgio Cortiana, and Jeanette Miriam Lorenz. Enhancing the scalability of classical surrogates for real-world quantum machine learning applications, 2025. URL:https://arxiv.org/abs/2508.06131,arXiv:2508.06131

  37. [37]

    Marshall, Riccardo Molteni, and Vedran Dunjko

    Sofiene Jerbi, Casper Gyurik, Simon C. Marshall, Riccardo Molteni, and Vedran Dunjko. Shadows of quantum machine learning.Nature Communications, 15(1), July 2024. URL:http://dx.doi.org/ 10.1038/s41467-024-49877-8,doi:10.1038/s41467-024-49877-8

  38. [38]

    Tensor network surrogate models for variational quantum computation

    Ryo Watanabe, Dries Sels, and Joseph Tindall. Tensor network surrogate models for variational quantum computation, 2026. URL:https://arxiv.org/abs/2604.20180,arXiv:2604.20180

  39. [39]

    Local surrogates for quantum machine learning, 2025

    Sreeraj Rajindran Nair and Christopher Ferrie. Local surrogates for quantum machine learning, 2025. URL:https://arxiv.org/abs/2506.09425,arXiv:2506.09425

  40. [40]

    3 edition, 2025

    Christoph Molnar.Interpretable Machine Learning. 3 edition, 2025. URL:https://christophm. github.io/interpretable-ml-book

  41. [41]

    Longo, M

    Luca Longo, Mario Brcic, Federico Cabitza, Jaesik Choi, Roberto Confalonieri, Javier Del Ser, Ric- cardo Guidotti, Yoichi Hayashi, Francisco Herrera, Andreas Holzinger, Richard Jiang, Hassan Khos- ravi, Freddy Lecue, Gianclaudio Malgieri, Andr´ es P´ aez, Wojciech Samek, Johannes Schneider, Timo Speith, and Simone Stumpf. Explainable artificial intelligen...

  42. [42]

    Why Should I Trust You?

    Marco Tulio Ribeiro, Sameer Singh, and Carlos Guestrin. ”why should i trust you?”: Explaining the predictions of any classifier. InProceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD ’16, page 1135–1144, New York, NY, USA, 2016. Association for Computing Machinery.doi:10.1145/2939672.2939778

  43. [43]

    On the interpretability of quantum neural networks.Quantum Machine Intelligence, 6(2), August 2024

    Lirand¨ e Pira and Chris Ferrie. On the interpretability of quantum neural networks.Quantum Machine Intelligence, 6(2), August 2024. URL:http://dx.doi.org/10.1007/s42484-024-00191-y,doi: 10.1007/s42484-024-00191-y

  44. [44]

    Rudolph, Armando Angrisani, Tyson Jones, M

    Sacha Lerch, Ricard Puig, Manuel S. Rudolph, Armando Angrisani, Tyson Jones, M. Cerezo, Supanut Thanasilp, and Zo¨ e Holmes. Efficient quantum-enhanced classical simulation for patches of quantum landscapes, 2024.arXiv:2411.19896

  45. [45]

    Rudolph, Thiparat Chotibut, Supanut Thanasilp, and Zo¨ e Holmes

    Hela Mhiri, Ricard Puig, Sacha Lerch, Manuel S. Rudolph, Thiparat Chotibut, Supanut Thanasilp, and Zo¨ e Holmes. A unifying account of warm start guarantees for patches of quantum landscapes,

  46. [46]

    URL:https://arxiv.org/abs/2502.07889,arXiv:2502.07889

  47. [47]

    Towards quantum machine learning with tensor networks

    William Huggins, Piyush Patil, Bradley Mitchell, K Birgitta Whaley, and E Miles Stoudenmire. To- wards quantum machine learning with tensor networks.Quantum Science and Technology, 4(2):024001, January 2019. URL:http://dx.doi.org/10.1088/2058-9565/aaea94,doi:10.1088/2058-9565/ aaea94

  48. [48]

    Deep tensor networks with matrix product operators.Quantum Machine In- telligence, 4(2), August 2022

    Bojan ˇZunkoviˇ c. Deep tensor networks with matrix product operators.Quantum Machine In- telligence, 4(2), August 2022. URL:http://dx.doi.org/10.1007/s42484-022-00081-1,doi: 10.1007/s42484-022-00081-1

  49. [49]

    Tensor networks for quantum machine learn- ing.Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences, 479(2275), July 2023

    Hans-Martin Rieser, Frank K¨ oster, and Arne Peter Raulf. Tensor networks for quantum machine learn- ing.Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences, 479(2275), July 2023. URL:http://dx.doi.org/10.1098/rspa.2023.0218,doi:10.1098/rspa.2023.0218

  50. [50]

    Hand-waving and interpretive dance: an introductory course on tensor networks.Journal of Physics A: Mathematical and Theoretical, 50(22):223001, May

    Jacob C Bridgeman and Christopher T Chubb. Hand-waving and interpretive dance: an introductory course on tensor networks.Journal of Physics A: Mathematical and Theoretical, 50(22):223001, May

  51. [51]

    URL:http://dx.doi.org/10.1088/1751-8121/aa6dc3,doi:10.1088/1751-8121/aa6dc3

  52. [52]

    J. Eisert. Entanglement and tensor network states, 2013. URL:https://arxiv.org/abs/1308.3318, arXiv:1308.3318

  53. [53]

    Oseledets , title =

    Ivan V. Oseledets. Tensor-train decomposition.SIAM Journal on Scientific Computing, 33(5):2295– 2317, 2011.doi:10.1137/090752286

  54. [54]

    Perez-Garcia, F

    D. Perez-Garcia, F. Verstraete, M. M. Wolf, and J. I. Cirac. Matrix product state representations,

  55. [55]

    URL:https://arxiv.org/abs/quant-ph/0608197,arXiv:quant-ph/0608197

  56. [56]

    Understanding Machine Learning

    Shai Shalev-Shwartz and Shai Ben-David.Understanding Machine Learning: From Theory to Algo- rithms. Cambridge University Press, 2014.doi:10.1017/CBO9781107298019

  57. [57]

    Chebyshev approximation and composition of functions in matrix product states for quantum-inspired numerical analysis, 2025

    Juan Jos´ e Rodr´ ıguez-Aldavero, Paula Garc´ ıa-Molina, Luca Tagliacozzo, and Juan Jos´ e Garc´ ıa-Ripoll. Chebyshev approximation and composition of functions in matrix product states for quantum-inspired numerical analysis, 2025. URL:https://arxiv.org/abs/2407.09609,arXiv:2407.09609

  58. [58]

    V.N. Vapnik. An overview of statistical learning theory.IEEE Transactions on Neural Networks, 10(5):988–999, 1999.doi:10.1109/72.788640

  59. [59]

    Lower and upper bounds on the pseudo-dimension of tensor network models

    Behnoush Khavari and Guillaume Rabusseau. Lower and upper bounds on the pseudo-dimension of tensor network models. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman 24 Vaughan, editors,Advances in Neural Information Processing Systems, volume 34, pages 10931–10943. Curran Associates, Inc., 2021. URL:https://proceedings.neurips.cc/paper...

  60. [60]

    Learnability, stability and uniform convergence.Journal of Machine Learning Research, 11(90):2635–2670, 2010

    Shai Shalev-Shwartz, Ohad Shamir, Nathan Srebro, and Karthik Sridharan. Learnability, stability and uniform convergence.Journal of Machine Learning Research, 11(90):2635–2670, 2010. URL: http://jmlr.org/papers/v11/shalev-shwartz10a.html

  61. [61]

    MIT Press, 2 edition, 2018

    Mehryar Mohri, Afshin Rostamizadeh, and Ameet Talwalkar.Foundations of Machine Learning. MIT Press, 2 edition, 2018

  62. [62]

    Springer: New York, 1999

    Vladimir Vapnik.The Nature of Statistical Learning Theory. Springer: New York, 1999

  63. [63]

    A variational approach of the rank function.TOP, 21(2):207–240, 2013.doi:10.1007/s11750-012-0234-4

    Jean-Baptiste Hiriart-Urruty and Hoang Tuy Le. A variational approach of the rank function.TOP, 21(2):207–240, 2013.doi:10.1007/s11750-012-0234-4

  64. [64]

    Anselm Blumer, Andrzej Ehrenfeucht, David Haussler, and Manfred K. Warmuth. Learnability and the Vapnik-Chervonenkis dimension.Journal of the ACM, 36(4):929–965, 1989.doi:10.1145/76359. 76371

  65. [65]

    UCI Machine Learning Repository (2012)

    Volker Lohweg. Banknote Authentication. UCI Machine Learning Repository, 2012. DOI: https://doi.org/10.24432/C55P57

  66. [66]

    N. Sauer. On the density of families of sets.Journal of Combinatorial Theory, Series A, 13(1):145–147, 1972.doi:10.1016/0097-3165(72)90019-2. A Appendix: VC Dimension Union Bound We provide a self-contained proof of the union bound for VC dimension used in Lemma 7.1. Lemma A.1(VC Dimension of Union).LetC 1,C 2 be set families over a domainXwithVCdim(C 1) ...