pith. machine review for the scientific record. sign in

arxiv: 2604.13096 · v1 · submitted 2026-04-09 · 🧮 math.GM

Recognition: unknown

Complexity scaling and optimal policy degeneracy in quantum reinforcement learning via analytically solvable unitary-control-then-measure models

Authors on Pith no claims yet

Pith reviewed 2026-05-10 17:04 UTC · model grok-4.3

classification 🧮 math.GM
keywords quantum reinforcement learningunitary controlprojective measurementexpected returncomputational complexitypolicy degeneracyquantum Zeno effectMarkov decision processes
0
0 comments X

The pith

Quantum RL with unitary control and measurement reduces expected return complexity from exponential to power-law scaling while revealing distinct optimal policy degeneracy patterns.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper constructs finite-dimensional quantum reinforcement learning models using a unitary-control-then-measure protocol that interleaves unitary transformations on a quantum state with projective measurements in a fixed basis. It derives exact closed-form expressions for trajectory probabilities, rewards, and expected returns in four specific cases: closed-chain and anti-periodic qubits, a qutrit ladder, and a four-level two-qubit system. These expressions allow the authors to demonstrate a two-level complexity reduction: first from path equivalence classes that group trajectories by state counts and transition frequencies, and second from the sparse transition graphs created by the constrained unitaries. The same models also show that optimal policies are unique in low dimensions with asymptotic behavior tied to the quantum Zeno effect, while the four-level case exhibits plateau quasi-degeneracy at long horizons and discrete degeneracy at critical parameter values.

Core claim

In these analytically solvable unitary-control-then-measure models of finite-horizon quantum Markov decision processes, the expected return admits exact closed-form expressions whose computational cost scales as O(N^I) rather than the nominal O(e^N), due to trajectory equivalence classes at the path level and sparsity of the allowed transition graph at the policy level; low-dimensional realizations possess unique optimal policies whose large-N limit is governed by the quantum Zeno effect, whereas the four-level system displays both plateau-type quasi-degeneracy for large horizons and genuine discrete degeneracy at critical energy parameters.

What carries the argument

The unitary-control-then-measure protocol, which applies a unitary to the quantum state and immediately follows it with a projective measurement onto a fixed reference basis, enabling closed-form trajectory probabilities and expected-return sums.

If this is right

  • In the qubit and qutrit models, optimal policies remain unique for every finite horizon, with their large-N behavior fixed by the quantum Zeno effect.
  • The four-level two-qubit model exhibits both continuous plateau quasi-degeneracy at large N and isolated points of discrete policy degeneracy at specific critical energy values.
  • The two-level complexity reduction applies uniformly across all four realizations, converting the sum over exponentially many trajectories into a polynomial-time sum over equivalence classes.
  • The sparsity of the transition graph induced by unitary constraints is what produces the second, policy-level reduction beyond the combinatorial equivalence at the trajectory level.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • If the same equivalence-class counting extends to higher-dimensional or continuous-variable systems, then practical quantum RL algorithms could evaluate returns for horizons previously considered intractable.
  • The observed discrete degeneracy at critical parameters suggests that small changes in Hamiltonian energies could be used to engineer policy multiplicity or uniqueness on demand.
  • Because the models are fully solvable, they provide exact benchmarks against which approximate quantum RL methods on larger systems can be validated.

Load-bearing premise

The assumption that the chosen finite-dimensional unitary controls and fixed-basis projective measurements permit exact closed-form derivations while still capturing the essential complexity and degeneracy features of broader quantum reinforcement learning.

What would settle it

Direct numerical enumeration of all possible trajectories for the four-level model at increasing horizon lengths N, checking whether the time to compute the expected return grows as a power of N rather than exponentially.

Figures

Figures reproduced from arXiv: 2604.13096 by Alessandro Michelangeli, Andrea Cintio, Dmitrii Tsutskov.

Figure 1
Figure 1. Figure 1: Numerical optimisation of the Expected Return (4.19) the additional positive return from these terms (specifically from the (p−1)x+ parts), the optimization pushes x− to be strictly positive. • The penalty limits x−: the reward kernel inside the sum contains the negative term −(N + 1 − p)x−. Since this coefficient scales as O(N), x− must scale as O(1/N) to keep the penalty finite (order O(1)). Consequently… view at source ↗
Figure 2
Figure 2. Figure 2: Numerical optimisation of the Expected Return (5.7) Therefore, the expected return for this model is J(π) = X N p=1 min{p X ,N+1−p} c=1  p − 1 c − 1  ·  N − p c − 1  [PITH_FULL_IMAGE:figures/full_fig_p011_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: Left: the directed graph G3 on node set {|0⟩, |1⟩, |2⟩} with all six off-diagonal edges, whose incidence matrix has rank 3 − 1 = 2 [4, Theo￾rem 2.3]. Right: the reduced transition graph Ge3, obtained from G3 by re￾taining only the three edges |i⟩ → |j⟩ with nonzero probability after imposing the forbidden-transition constraints (6.14); edge labels indicate the transi￾tion multiplicities aij from (6.16), fo… view at source ↗
Figure 4
Figure 4. Figure 4: Numerical optimisation of the Expected Return (6.31) The latter formula has a symmetry that is explicitly seen by restoring n1 from n0+n1+n2 = N + 1. One then re-writes Jε(x) = N X−1 n0=1 N X−n0 n2=1 min{n0X,n1,n2}−1 c=0  n0 − 1 c n1 − 1 c n2 − 1 c  ×  ε(n1 − n0) + (n0 + 2n2 − N − 2) x + 1 (1 − x) N−3c−2x 3c+2 , with n0 + n1 + n2 = N + 1 . (6.30) Now, swapping n0 ↔ n1 leaves the product of the bi… view at source ↗
Figure 5
Figure 5. Figure 5: Optimal policy (x ∗ , y∗ , z∗ ) for the expected return (6.34) as a func￾tion of the horizon N, for ε = 0.75 (left) and ε = 3.0 (right). The saturations y ∗ → 1 and z ∗ → 0 are interpreted in the text. dominates over reward: the gain from eliminating the feedback arc outweighs the foregone (n2 − 1)z term, since each feedback cycle would re-expose the system to the now-dominant ε-penalty at |0⟩. 7. A compar… view at source ↗
Figure 6
Figure 6. Figure 6: Comparison of computational time required to evaluate the ex￾pected return Jε(x) as a function of trajectory length N. The ‘Brute Force’ approach (red dashed line) scales exponentially as O(3N ), while the analytic formula (blue solid line) scales polynomially as O(N3 ). The sum 7.1 runs over the 3 N−1 possible configurations of the intermediate states, and is parametrised by the intermediate energy level … view at source ↗
Figure 7
Figure 7. Figure 7: Left: the directed graph G4 on node set {|0⟩, |1⟩, |1 ′ ⟩, |2⟩} with all twelve off-diagonal edges (i ̸= j), whose incidence matrix has rank 4 − 1 = 3 [4, Theorem 2.3]. Right: the reduced transition graph Ge4, obtained from G4 by retaining only the five edges |i⟩ → |j⟩ with nonzero probability after imposing the forbidden-transition constraints (8.24); edge labels indicate the transition multiplicities aij… view at source ↗
Figure 8
Figure 8. Figure 8: Numerical optimisation of the expected return (8.33) on a horizon of N = 8 long trajectories. The plots show the migration of the optimal policy inside the control space [0, 1]×[0, 1] as the energy penalties (ε, ε′ ) are varied. Remarkably, the linear dependence on the branching fluxes c01 and c01′ cancels out exactly (e.g., the energy cost of entering, traversing, and exiting the upper branch is balanced … view at source ↗
Figure 9
Figure 9. Figure 9: Numerical evidence of the emergence of a plateau-like region in the profile of the return function (8.33) around the point of optimal policy, as N increases [PITH_FULL_IMAGE:figures/full_fig_p029_9.png] view at source ↗
Figure 10
Figure 10. Figure 10: Emergence of degenerate optimal policies for the expected re￾turn (8.33) (horizon N = 8). A crossover of the global maximum (xmax, x′ max) between two distinct regions of the domain is observed when increasing ε from 2.47 to 2.48. Since the return Jε,1 is continuous in ε, this implies the existence of a critical value ε ∗ ∈ (2.47, 2.48) where two distinct optimal poli￾cies coexist with J max ε ∗,1 ≈ 0.644… view at source ↗
read the original abstract

We propose and analyse a class of analytically solvable models of quantum reinforcement learning (QRL), formulated as finite-horizon Markov decision processes in finite-dimensional Hilbert spaces. The models are built around a `unitary-control-then-measure' protocol, in which a learning agent applies unitary transformations to a quantum state and interleaves each control step with a projective measurement onto a prescribed reference basis. Exact closed-form expressions for trajectory probabilities, rewards, and the expected return are derived for four concrete realisations: a closed-chain and an anti-periodic qubit implementation, a qutrit model with ladder coupling, and a four-level two-qubit system. Two structural features of these QRL protocols are rigorously analysed. First, we identify and quantify a two-level reduction in the computational complexity of the expected return, from the nominally exponential $O(e^N)$ scaling in the trajectory length~$N$ to an explicit power-law $O(N^{\mathcal{I}})$: a trajectory-based level, arising from equivalence classes of paths sharing the same unordered state counts and transition frequencies, and a policy-based level, arising from the sparsity of the transition graph enforced by constrained unitary actions. Second, we characterise the degeneracy of optimal policies. The low-dimensional models exhibit unique optima whose asymptotic behaviour with~$N$ is governed by the quantum Zeno effect, while the four-level system displays both plateau-type quasi-degeneracy at large horizons and genuine discrete degeneracy at critical energy parameters -- phenomena with no counterpart in the measurement-free quantum optimal control landscape.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 2 minor

Summary. The manuscript proposes and analyzes a class of analytically solvable quantum reinforcement learning models formulated as finite-horizon MDPs in finite-dimensional Hilbert spaces. These are constructed around a unitary-control-then-measure protocol, with exact closed-form expressions derived for trajectory probabilities, rewards, and expected returns in four explicit low-dimensional realizations (closed-chain qubit, anti-periodic qubit, qutrit ladder, and four-level two-qubit systems). The central results are a two-level complexity reduction for the expected return—from nominal O(e^N) to O(N^I) scaling, arising from trajectory equivalence classes (unordered state counts and transition frequencies) plus unitary-enforced transition-graph sparsity—and a characterization of optimal-policy degeneracy, including unique optima with quantum-Zeno asymptotics in the qubit/qutrit cases and both plateau quasi-degeneracy and discrete degeneracy at critical energies in the four-level model.

Significance. If the closed-form derivations and scaling claims hold, the work supplies rare exactly solvable benchmarks in QRL that quantify how unitary constraints and projective measurements induce polynomial complexity and specific degeneracy structures absent from measurement-free quantum control. The explicit model constructions and analytical results constitute a clear strength, enabling reproducible verification and potential generalization of the equivalence-class and sparsity mechanisms.

major comments (2)
  1. [§4] §4 (complexity reduction): the reduction from O(e^N) to O(N^I) is asserted to follow from equivalence classes of trajectories and sparsity of the transition graph; however, the manuscript must explicitly compute or bound the exponent I for each of the four models as a function of N and Hilbert-space dimension, including verification that the number of distinct (state-count, transition-frequency) classes remains polynomial for arbitrary N rather than merely for the small-N cases examined.
  2. [§5.2] §5.2 (four-level degeneracy): the distinction between plateau-type quasi-degeneracy at large N and genuine discrete degeneracy at critical energy parameters is central to the claim of phenomena with no classical counterpart; the analysis should supply the explicit algebraic condition on the energy parameters that produces the discrete degeneracy and demonstrate that it is not an artifact of the finite-N truncation.
minor comments (2)
  1. The notation for the polynomial degree I is introduced without a dedicated definition or table summarizing its value per model; a compact table or explicit formula in the main text would improve readability.
  2. Figure captions for the degeneracy plots should state the precise numerical tolerances used to identify 'plateau' versus 'discrete' degeneracy.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the careful reading, positive assessment, and recommendation for minor revision. We address each major comment below and will incorporate the requested clarifications and extensions into the revised manuscript.

read point-by-point responses
  1. Referee: [§4] §4 (complexity reduction): the reduction from O(e^N) to O(N^I) is asserted to follow from equivalence classes of trajectories and sparsity of the transition graph; however, the manuscript must explicitly compute or bound the exponent I for each of the four models as a function of N and Hilbert-space dimension, including verification that the number of distinct (state-count, transition-frequency) classes remains polynomial for arbitrary N rather than merely for the small-N cases examined.

    Authors: We agree that explicit per-model expressions for the exponent I are needed to substantiate the claim. In the revision we will derive and state the precise polynomial degree I for each of the four realizations (closed-chain qubit, anti-periodic qubit, qutrit ladder, four-level two-qubit), expressing I explicitly in terms of horizon N and Hilbert-space dimension d. We will also supply a general counting argument showing that the number of distinct equivalence classes—unordered state-visit multisets and transition-frequency vectors compatible with the unitary transition graph—is bounded by a polynomial of degree at most d−1 in N for arbitrary N, thereby confirming the O(N^I) scaling holds beyond the small-N examples. revision: yes

  2. Referee: [§5.2] §5.2 (four-level degeneracy): the distinction between plateau-type quasi-degeneracy at large N and genuine discrete degeneracy at critical energy parameters is central to the claim of phenomena with no classical counterpart; the analysis should supply the explicit algebraic condition on the energy parameters that produces the discrete degeneracy and demonstrate that it is not an artifact of the finite-N truncation.

    Authors: We thank the referee for this observation. The discrete degeneracy occurs precisely when the energy parameters satisfy a specific algebraic relation obtained by requiring that two distinct policy parameters yield identical expected returns; this relation is independent of N because the return expression factors into an N-independent prefactor times a polynomial whose roots determine the critical energies. In the revision we will state the explicit algebraic condition and prove that the same roots persist for all N by direct inspection of the closed-form return, confirming the degeneracy is not an artifact of finite-horizon truncation. revision: yes

Circularity Check

0 steps flagged

No significant circularity; derivations are direct consequences of model definitions

full rationale

The paper defines a specific 'unitary-control-then-measure' protocol in finite-dimensional Hilbert spaces and constructs four concrete low-dimensional models (closed-chain qubit, anti-periodic qubit, qutrit ladder, four-level two-qubit). It then derives exact closed-form expressions for trajectory probabilities, rewards, and expected returns directly from these definitions. The claimed two-level complexity reduction to O(N^I) follows mathematically from grouping trajectories into equivalence classes by unordered state counts and transition frequencies (a combinatorial consequence of the finite state space and measurement basis) plus the sparsity of the transition graph enforced by the constrained unitaries. No parameters are fitted to data and then relabeled as predictions; no self-citations or uniqueness theorems from prior author work are invoked as load-bearing steps; and no ansatz is smuggled in. The results are therefore self-contained analytical consequences rather than reductions to the inputs by construction.

Axiom & Free-Parameter Ledger

1 free parameters · 3 axioms · 0 invented entities

The models rest on standard quantum mechanics and MDP concepts; no new physical entities are postulated. One potential free parameter (critical energy values) is mentioned only in the context of degeneracy points.

free parameters (1)
  • critical energy parameters
    Referenced as the values at which the four-level system exhibits genuine discrete degeneracy; their explicit form is not given in the abstract.
axioms (3)
  • standard math Unitary operators on finite-dimensional Hilbert spaces preserve state norms and enable coherent control
    Invoked throughout the unitary-control-then-measure protocol.
  • standard math Projective measurements onto a fixed reference basis produce probabilistic outcomes according to Born rule
    Core of the interleaving measurement step.
  • domain assumption Finite-horizon Markov decision process formulation applies to the quantum state evolution
    Used to define rewards and expected return.

pith-pipeline@v0.9.0 · 5583 in / 1524 out tokens · 105994 ms · 2026-05-10T17:04:50.361582+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

29 extracted references · 1 canonical work pages

  1. [1]

    Aharonov, L

    Y. Aharonov, L. Davidovich, and N. Zagury , Quantum random walks , Phys. Rev. A, 48 (1993), pp. 1687--1690

  2. [2]

    Arulkumaran, M

    K. Arulkumaran, M. P. Deisenroth, M. Brundage, and A. A. Bharath , Deep reinforcement learning: A brief survey , IEEE signal processing magazine, 34 (2017), pp. 26--38

  3. [3]

    Attal, F

    S. Attal, F. Petruccione, C. Sabot, and I. Sinayskiy , Open Quantum Random Walks , J. Stat. Phys., 147 (2012), pp. 832--852

  4. [4]

    R. B. Bapat , Graphs and Matrices , Universitext , Springer, London, 2 ed., 2014

  5. [5]

    A. Barr, W. Gispen, and A. Lamacraft , Quantum Ground States from Reinforcement Learning , in Proceedings of Machine Learning Research , vol. 107, PMLR, 2020, pp. 635--653

  6. [6]

    Bertsekas , Reinforcement learning and optimal control , vol

    D. Bertsekas , Reinforcement learning and optimal control , vol. 1, Athena Scientific, 2019

  7. [7]

    height 2pt depth -1.6pt width 23pt, A course in reinforcement learning , Athena Scientific, 2024

  8. [8]

    H. J. Briegel and G. De las Cuevas , Projective simulation for artificial intelligence , Sci. Rep., 2 (2012), p. 400

  9. [9]

    Clarke and F

    J. Clarke and F. K. Wilhelm , Superconducting quantum bits , Nature, 453 (2008), pp. pages 1031--1042

  10. [10]

    Cohen-Tannoudji, B

    C. Cohen-Tannoudji, B. Diu, and F. Lalo \"e , Quantum mechanics , Wiley, New York, NY, 1977. Trans. of : M \'e canique quantique. Paris : Hermann, 1973

  11. [11]

    D. Dong, C. Chen, H. Li, and T.-J. Tarn , Quantum Reinforcement Learning , IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics), 38 (2008), pp. 1207--1220

  12. [12]

    1207--1220

    height 2pt depth -1.6pt width 23pt, Quantum reinforcement learning , IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics), 38 (2008), pp. 1207--1220

  13. [13]

    Facchi and S

    P. Facchi and S. Pascazio , Quantum Zeno dynamics: mathematical and physical aspects , Journal of Physics A: Mathematical and Theoretical, 41 (2008), p. 493001

  14. [14]

    Hsieh and H

    M. Hsieh and H. Rabitz , Optimal control landscape for the generation of unitary transformations , Phys. Rev. A, 77 (2008), p. 042306

  15. [15]

    Hsieh, R

    M. Hsieh, R. Wu, H. Rabitz, and D. Lidar , Optimal control landscape for the generation of unitary transformations with constrained dynamics , Phys. Rev. A, 81 (2010), p. 062352

  16. [16]

    L. P. Kaelbling, M. L. Littman, and A. W. Moore , Reinforcement learning: A survey , Journal of artificial intelligence research, 4 (1996), pp. 237--285

  17. [17]

    Leibfried, R

    D. Leibfried, R. Blatt, C. Monroe, and D. Wineland , Quantum dynamics of single trapped ions , Rev. Mod. Phys., 75 (2003), pp. 281--324

  18. [18]

    A survey on quantum reinforcement learning,

    N. Meyer, C. Ufrecht, M. Periyasamy, D. D. Scherer, A. Plinge, and C. Mutschler , A Survey on Quantum Reinforcement Learning , arXiv:2211.03464 (2022)

  19. [19]

    G. D. Paparo, V. Dunjko, A. Makmal, M. A. Martin-Delgado, and H. J. Briegel , Quantum speed-up for active learning agents , Phys. Rev. X, 4 (2014), p. 031002

  20. [20]

    M. L. Puterman , Markov decision processes: discrete stochastic dynamic programming , John Wiley & Sons, 2014

  21. [21]

    S. M. Reimann and M. Manninen , Electronic structure of quantum dots , Rev. Mod. Phys., 74 (2002), pp. 1283--1342

  22. [22]

    Schenk, E

    M. Schenk, E. F. Combarro, M. Grossi, V. Kain, K. S. B. Li, M.-M. Popa, and S. Vallecorsa , Hybrid actor-critic algorithm for quantum reinforcement learning at cern beam lines , Quantum Science and Technology, 9 (2024), p. 025012

  23. [23]

    Sugny and C

    D. Sugny and C. Kontz , Optimal control of a three-level quantum system by laser fields plus von Neumann measurements , Phys. Rev. A, 77 (2008), p. 063420

  24. [24]

    R. S. Sutton and A. G. Barto , Reinforcement Learning: An Introduction , The MIT Press, second ed., 2018

  25. [25]

    Szepesv \'a ri , Algorithms for reinforcement learning , Springer nature, 2022

    C. Szepesv \'a ri , Algorithms for reinforcement learning , Springer nature, 2022

  26. [26]

    S. E. Venegas-Andraca , Quantum walks: a comprehensive review , Quantum Inf. Process., 11 (2012), pp. 1015--1106

  27. [27]

    Volkov, A

    B. Volkov, A. Myachkova, and A. Pechen , Phenomenon of a stronger trapping behavior in -type quantum systems with symmetry , Phys. Rev. A, 111 (2025), p. 022617

  28. [28]

    Wendin , Quantum information processing with superconducting circuits: a review , Reports on Progress in Physics, 80 (2017), p

    G. Wendin , Quantum information processing with superconducting circuits: a review , Reports on Progress in Physics, 80 (2017), p. 106001

  29. [29]

    S. Wu, S. Jin, D. Wen, D. Han, and X. Wang , Quantum reinforcement learning in continuous action space , Quantum, 9 (2025), p. 1660