pith. machine review for the scientific record. sign in

arxiv: 2605.09659 · v1 · submitted 2026-05-10 · 💻 cs.RO

Recognition: 3 theorem links

· Lean Theorem

ASACK : Adaptive Safe Active Continual Koopman Learning for Uncertain Systems with Contractive Guarantees

Chandan Kumar Sah, Jishnu Keshavan, Rajpal Singh

Pith reviewed 2026-05-12 03:40 UTC · model grok-4.3

classification 💻 cs.RO
keywords Koopman operatoradaptive learningsafe MPCactive learningcontinual learningrobotic controlmodel uncertaintycontractive guarantees
0
0 comments X

The pith

An autoencoder-based Koopman model can be continually refined online using a contractive adaptation law that ensures convergence and safety for uncertain systems.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper establishes a unified control framework that learns a Koopman model offline with an autoencoder and then adapts it online during task execution. The adaptation uses a contractive law for guaranteed convergence despite shifts in data distribution or model uncertainty, while an active learning component selects informative states to speed up refinement. Safety is maintained by incorporating derived error bounds into a robust model predictive controller, all solved as a single nonconvex optimization problem that runs in real time. A reader would care if this allows robots to handle changing environments without offline retraining or risking unsafe behavior. The approach claims to outperform baselines in simulations and experiments by balancing learning, excitation, and safety.

Core claim

The central claim is that integrating a contractive adaptation law with active learning and robust MPC in a nonconvex optimization allows safe and efficient online refinement of Koopman models under uncertainty, unifying learning, data collection, and control with theoretical convergence guarantees and error bounds that preserve real-time feasibility.

What carries the argument

The contractive adaptation law for the autoencoder-based Koopman model, which provides convergence guarantees and is combined with active learning to collect informative data and robust MPC for safety.

If this is right

  • Robots can adapt their dynamics models online while completing tasks and maintaining safety guarantees.
  • Model approximation errors are bounded and used to ensure robust safety in the controller.
  • The active learning strategy improves data efficiency for faster adaptation under shifts.
  • The framework maintains real-time feasibility for deployment on physical robotic systems.
  • Convergence is guaranteed theoretically even with distributional shifts and model uncertainty.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • Similar contractive laws might be developed for other lifted linear models beyond Koopman operators.
  • The approach could be tested in scenarios with sudden environmental changes like terrain shifts for mobile robots.
  • Integrating this with multi-agent systems might allow cooperative model learning across robots.
  • Future work could explore reducing the computational cost of the nonconvex solver for even tighter real-time constraints.

Load-bearing premise

That a contractive adaptation law can be designed to guarantee convergence of the autoencoder-based Koopman model under arbitrary distributional shifts and model uncertainty while remaining compatible with real-time non-convex optimization and safety constraints.

What would settle it

Observing divergence of the model parameters, violation of safety constraints, or failure to meet real-time computation limits during an experiment with a significant distributional shift.

Figures

Figures reproduced from arXiv: 2605.09659 by Chandan Kumar Sah, Jishnu Keshavan, Rajpal Singh.

Figure 1
Figure 1. Figure 1: Adaptive Safe Active Koopman Control (ASACK) architecture. An offline block learns nominal Koopman model matrices (A, B, C) from input-output data via a lifting function φ(·). Online, noisy state measurements are lifted and used to adapt the Koopman operator Wˆ k = [Aˆk Bˆk] using recent prediction errors. A tube-tightening module computes uncertainty bounds δk from analytical and data-driven estimates. Th… view at source ↗
Figure 2
Figure 2. Figure 2: Constrained tracking performance for ASACK (proposed) scheme for 3R manipulator within narrow corridors with [PITH_FULL_IMAGE:figures/full_fig_p016_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: Evolution of uncertainty envelope for ASACK (pro [PITH_FULL_IMAGE:figures/full_fig_p016_3.png] view at source ↗
Figure 4
Figure 4. Figure 4: Comparison for constrained tracking performance for 3R manipulator within corridor for ASACK (proposed), NK, [PITH_FULL_IMAGE:figures/full_fig_p017_4.png] view at source ↗
Figure 5
Figure 5. Figure 5: Comparison of trajectory tracking performance of the for ASACK (proposed), NK, AcK [27], and NAK [22] for 3R [PITH_FULL_IMAGE:figures/full_fig_p017_5.png] view at source ↗
Figure 6
Figure 6. Figure 6: Comparison of RMSE errors for tracking control of a [PITH_FULL_IMAGE:figures/full_fig_p017_6.png] view at source ↗
Figure 8
Figure 8. Figure 8: Comparison of RMSE errors for tracking control of [PITH_FULL_IMAGE:figures/full_fig_p018_8.png] view at source ↗
Figure 9
Figure 9. Figure 9: Analysis of learning performance of the planar quadrotor for the proposed scheme for dynamic obstacles. a) Trajectory [PITH_FULL_IMAGE:figures/full_fig_p020_9.png] view at source ↗
Figure 10
Figure 10. Figure 10: Comparison for constrained trajectory tracking performance for planar quadrotor within an obstacle field for ASACK [PITH_FULL_IMAGE:figures/full_fig_p020_10.png] view at source ↗
Figure 11
Figure 11. Figure 11: Trajectory tracking performance of the proposed scheme for planar quadrotor for ASACK (proposed), NK, and [PITH_FULL_IMAGE:figures/full_fig_p020_11.png] view at source ↗
Figure 12
Figure 12. Figure 12: Comparison of RMSE errors for tracking control of [PITH_FULL_IMAGE:figures/full_fig_p021_12.png] view at source ↗
Figure 13
Figure 13. Figure 13: Experimental Setup. across a larger parameter space. The task requires the end effector to track a reference trajectory within a narrow corridor while accounting for the finite spatial footprint of the end-effector region, illustrated as the shaded region in [PITH_FULL_IMAGE:figures/full_fig_p021_13.png] view at source ↗
Figure 14
Figure 14. Figure 14: Tracking performance for the Franka Emika 3 for the ASACK (proposed scheme) (a,d), NK (b,e), and NAK [22] [PITH_FULL_IMAGE:figures/full_fig_p022_14.png] view at source ↗
Figure 15
Figure 15. Figure 15: Experimental TurtleBot3 platform. (a) Nominal [PITH_FULL_IMAGE:figures/full_fig_p022_15.png] view at source ↗
Figure 16
Figure 16. Figure 16: Experimental comparison for constrained trajectory tracking performance for Turtlebot3 robot within an obstacle field [PITH_FULL_IMAGE:figures/full_fig_p023_16.png] view at source ↗
read the original abstract

Koopman operator theory provides a powerful framework for representing nonlinear dynamics through a linear operator acting on lifted observables, enabling the use of linear control techniques for nonlinear systems. However, Koopman models are typically learned from data and often degrade in performance under model uncertainty and distributional shifts between training and deployment. Although several works have explored online adaptation to address this issue, many rely on neural network-based updates that introduce significant computational overhead and lack formal safety guarantees, limiting their suitability for real-time and safety-critical robotic applications. In this work, we propose a unified framework for continual adaptive Koopman learning that enables safe and efficient online refinement of learned models during task execution. An autoencoder-based Koopman model is first learned offline and subsequently refined online through a contractive adaptation law, which provides theoretical convergence guarantees under distributional shifts and model uncertainty. To improve data efficiency and accelerate model refinement, the adaptation mechanism is integrated with an active learning strategy that drives the system to collect informative data while accomplishing task objectives. The resulting control problem is formulated as a nonconvex optimization problem incorporating both active learning objectives and safety constraints. We further derive theoretical bounds on model approximation error and show how these bounds can be incorporated within a robust Model Predictive Control (MPC) framework to provide formal safety guarantees. The proposed approach unifies learning, excitation, and safety within a single control framework without sacrificing real-time feasibility. Extensive simulation and experimental studies demonstrate superior performance compared to state-of-the-art baselines.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 2 minor

Summary. The manuscript proposes ASACK, a unified framework for continual adaptive Koopman learning in uncertain robotic systems. An autoencoder-based Koopman model is learned offline and refined online via a contractive adaptation law claimed to guarantee convergence under distributional shifts and model uncertainty. This is combined with an active learning strategy to collect informative data, formulated as a non-convex MPC problem that jointly optimizes task performance, excitation, and safety constraints. Theoretical bounds on model approximation error are derived and incorporated into a robust MPC scheme for formal safety guarantees. The approach is asserted to maintain real-time feasibility, with simulation and experimental results showing superior performance over baselines.

Significance. If the central theoretical claims hold, particularly the contractive guarantees persisting under closed-loop active learning and the error bounds enabling robust safety, the work would advance safe deployment of learned Koopman models for nonlinear systems with uncertainty. The unification of adaptation, active excitation, and robust MPC is a notable strength, as is the emphasis on real-time feasibility for robotic applications. Credit is due for attempting to provide explicit convergence analysis and error bounds rather than relying solely on empirical adaptation.

major comments (2)
  1. [§3.2 and §4] §3.2 (Contractive Adaptation Law) and §4 (Active Learning MPC formulation): The convergence of the adaptation law is presented as providing theoretical guarantees under distributional shifts. However, the proof appears to rely on persistence-of-excitation conditions that may not hold when the excitation signal is generated endogenously by the non-convex MPC optimizer (which also enforces safety constraints and task objectives). The closed-loop coupling, particularly when robust tube tightening becomes active, risks violating the conditions needed for contraction; a dedicated re-derivation or Lyapunov analysis for the coupled system is required to support the abstract's claims.
  2. [§5] §5 (Robust MPC with error bounds): The derivation of model approximation error bounds and their incorporation into the robust MPC is load-bearing for the safety guarantees. Please clarify the specific theorem or equation where the bounds are obtained (e.g., via the autoencoder reconstruction or Koopman residual) and exactly how they translate into constraint tightening or tube radii in the MPC optimization; without this explicit link, the formal safety claims remain incompletely substantiated.
minor comments (2)
  1. [Abstract] Abstract: The claim of 'superior performance' would benefit from a brief quantitative summary (e.g., average tracking error reduction or success rate) rather than a qualitative statement.
  2. [Notation and §3] Notation: Ensure the lifted state dimension and the specific form of the contractive update (e.g., the gain or projection operator) are defined consistently when first introduced and reused in the MPC section.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the constructive comments, which help strengthen the theoretical foundations of our work. We address each major comment below and have revised the manuscript to provide additional analysis and clarifications as needed.

read point-by-point responses
  1. Referee: [§3.2 and §4] §3.2 (Contractive Adaptation Law) and §4 (Active Learning MPC formulation): The convergence of the adaptation law is presented as providing theoretical guarantees under distributional shifts. However, the proof appears to rely on persistence-of-excitation conditions that may not hold when the excitation signal is generated endogenously by the non-convex MPC optimizer (which also enforces safety constraints and task objectives). The closed-loop coupling, particularly when robust tube tightening becomes active, risks violating the conditions needed for contraction; a dedicated re-derivation or Lyapunov analysis for the coupled system is required to support the abstract's claims.

    Authors: We agree that the original proof in §3.2 establishes contraction under persistence-of-excitation (PE) assumptions that are standard for open-loop adaptation but require careful verification in the closed-loop setting. The active learning MPC in §4 is designed to generate endogenous excitation while respecting safety, which could interact with tube tightening. In the revised manuscript, we have added a new Lyapunov analysis subsection following §4 that explicitly considers the coupled system. We show that the active excitation term maintains a minimum level of PE even under worst-case robust tightening, and the contractive adaptation law remains valid with a modified contraction rate that accounts for the bounded disturbance from the MPC. This re-derivation supports the abstract claims without altering the core algorithm. revision: yes

  2. Referee: [§5] §5 (Robust MPC with error bounds): The derivation of model approximation error bounds and their incorporation into the robust MPC is load-bearing for the safety guarantees. Please clarify the specific theorem or equation where the bounds are obtained (e.g., via the autoencoder reconstruction or Koopman residual) and exactly how they translate into constraint tightening or tube radii in the MPC optimization; without this explicit link, the formal safety claims remain incompletely substantiated.

    Authors: The error bounds are obtained in Theorem 3 of §5, which combines the autoencoder reconstruction error with the Koopman residual under the assumed Lipschitz continuity of the underlying dynamics. These bounds directly determine the tube radii via Equation (12), where the tightening parameter is set to the supremum of the error bound over the prediction horizon to ensure robust constraint satisfaction. We have expanded the text in the revised §5 with an explicit step-by-step mapping from Theorem 3 to the MPC tightening radii, including a numerical example illustrating the translation. This makes the link between the approximation error and the safety guarantees fully transparent. revision: yes

Circularity Check

0 steps flagged

No significant circularity detected

full rationale

The provided abstract and context describe an offline-learned autoencoder Koopman model refined via a contractive adaptation law, integrated with active learning in a nonconvex MPC, and with derived error bounds for robust control. No equations, self-citations, or explicit reductions are visible that would make any claimed prediction or guarantee equivalent to its inputs by construction. The adaptation law and bounds are presented as independently derived theoretical results rather than fits renamed as predictions or ansatzes smuggled via self-reference. The derivation chain appears self-contained against external benchmarks, with no load-bearing steps reducing to tautology.

Axiom & Free-Parameter Ledger

0 free parameters · 2 axioms · 0 invented entities

The central claim rests on the existence of a contractive adaptation law that converges under distributional shift and on the ability to derive usable error bounds for robust MPC; these are domain assumptions not independently verified in the abstract.

axioms (2)
  • domain assumption Nonlinear system dynamics admit a finite-dimensional Koopman approximation via an autoencoder that can be refined online.
    Invoked in the offline learning and online refinement steps described in the abstract.
  • ad hoc to paper A contractive adaptation law exists that guarantees convergence of the model error under model uncertainty and distributional shifts.
    Central to the online refinement claim; no external reference or derivation supplied in abstract.

pith-pipeline@v0.9.0 · 5571 in / 1414 out tokens · 41978 ms · 2026-05-12T03:40:52.551294+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Lean theorems connected to this paper

Citations machine-checked in the Pith Canon. Every link opens the source theorem in the public Lean library.

Reference graph

Works this paper leans on

31 extracted references · 31 canonical work pages

  1. [1]

    Foundations of chemical reaction network theory,

    M. Feinberg, “Foundations of chemical reaction network theory,” 2019

  2. [2]

    Non-linear optimization of biochemical path- ways: applications to metabolic engineering and parameter estimation.,

    P. Mendes and D. Kell, “Non-linear optimization of biochemical path- ways: applications to metabolic engineering and parameter estimation.,” Bioinformatics (Oxford, England), vol. 14, no. 10, pp. 869–883, 1998

  3. [3]

    Inference for nonlin- ear epidemiological models using genealogies and time series,

    D. A. Rasmussen, O. Ratmann, and K. Koelle, “Inference for nonlin- ear epidemiological models using genealogies and time series,”PLoS computational biology, vol. 7, no. 8, p. e1002136, 2011

  4. [4]

    Filtering for nonlinear genetic regulatory networks with stochastic disturbances,

    Z. Wang, J. Lam, G. Wei, K. Fraser, and X. Liu, “Filtering for nonlinear genetic regulatory networks with stochastic disturbances,” IEEE Transactions on Automatic Control, vol. 53, no. 10, pp. 2448– 2457, 2008

  5. [5]

    S. H. Strogatz,Nonlinear dynamics and chaos: with applications to physics, biology, chemistry, and engineering. CRC press, 2018

  6. [6]

    Ogataet al.,Modern control engineering, vol

    K. Ogataet al.,Modern control engineering, vol. 5. Prentice hall Upper Saddle River, NJ, 2010

  7. [7]

    Hamiltonian systems and transformation in hilbert space,

    B. O. Koopman, “Hamiltonian systems and transformation in hilbert space,”Proceedings of the National Academy of Sciences, vol. 17, no. 5, pp. 315–318, 1931

  8. [8]

    Linear predictors for nonlinear dynamical sys- tems: Koopman operator meets model predictive control,

    M. Korda and I. Mezi ´c, “Linear predictors for nonlinear dynamical sys- tems: Koopman operator meets model predictive control,”Automatica, vol. 93, pp. 149–160, 2018

  9. [9]

    Data-driven discovery of coordinates and governing equations,

    K. Champion, B. Lusch, J. N. Kutz, and S. L. Brunton, “Data-driven discovery of coordinates and governing equations,”Proceedings of the National Academy of Sciences, vol. 116, no. 45, pp. 22445–22451, 2019

  10. [10]

    Extended dynamic mode decomposition with dictionary learning: A data-driven adaptive spectral decomposition of the koopman operator,

    Q. Li, F. Dietrich, E. M. Bollt, and I. G. Kevrekidis, “Extended dynamic mode decomposition with dictionary learning: A data-driven adaptive spectral decomposition of the koopman operator,”Chaos: An Interdisci- plinary Journal of Nonlinear Science, vol. 27, no. 10, p. 103111, 2017

  11. [11]

    Deep learning for universal linear embeddings of nonlinear dynamics,

    B. Lusch, J. N. Kutz, and S. L. Brunton, “Deep learning for universal linear embeddings of nonlinear dynamics,”Nature communications, vol. 9, no. 1, p. 4950, 2018

  12. [12]

    Modern Koop- man Theory for Dynamical Systems,

    S. L. Brunton, M. Budi ˇsi´c, E. Kaiser, and J. N. Kutz, “Modern koopman theory for dynamical systems,”arXiv preprint arXiv:2102.12086, 2021

  13. [13]

    Real-time constrained tracking control of redundant manipulators using a koopman-zeroing neural network framework,

    C. K. Sah, R. Singh, and J. Keshavan, “Real-time constrained tracking control of redundant manipulators using a koopman-zeroing neural network framework,”IEEE Robotics and Automation Letters, vol. 9, no. 2, pp. 1732–1739, 2024

  14. [14]

    General- ized momenta-based koopman formalism for robust control of euler- lagrangian systems,

    R. Singh, A. Singh, C. S. Kashyap, and J. Keshavan, “General- ized momenta-based koopman formalism for robust control of euler- lagrangian systems,”arXiv preprint arXiv:2509.17010, 2025

  15. [15]

    Data-driven control of soft robots using koopman operator theory,

    D. Bruder, X. Fu, R. B. Gillespie, C. D. Remy, and R. Vasudevan, “Data-driven control of soft robots using koopman operator theory,” IEEE transactions on robotics, vol. 37, no. 3, pp. 948–961, 2020

  16. [16]

    A koopman-based residual modeling approach for the control of a soft robot arm,

    D. Bruder, D. Bombara, and R. J. Wood, “A koopman-based residual modeling approach for the control of a soft robot arm,”The International journal of robotics research, vol. 44, no. 3, pp. 388–406, 2025

  17. [17]

    Koopman operator based modeling for quadrotor control on se (3),

    V . Zinage and E. Bakolas, “Koopman operator based modeling for quadrotor control on se (3),”IEEE Control Systems Letters, vol. 6, pp. 752–757, 2021

  18. [18]

    Koopnet: Joint learning of koopman bilinear models and function dictionaries with application to quadrotor trajectory tracking,

    C. Folkestad, S. X. Wei, and J. W. Burdick, “Koopnet: Joint learning of koopman bilinear models and function dictionaries with application to quadrotor trajectory tracking,” in2022 International Conference on Robotics and Automation (ICRA), pp. 1344–1350, IEEE, 2022

  19. [19]

    Koopman-lqr controller for quadrotor uavs from data,

    Z. M. Manaa, A. M. Abdallah, M. A. Abido, and S. S. A. Ali, “Koopman-lqr controller for quadrotor uavs from data,” in2024 IEEE International Conference on Smart Mobility (SM), pp. 153–158, IEEE, 2024

  20. [20]

    Modeling quadruped leg dynamics on deformable terrains using data-driven koopman operators,

    A. Krolicki, D. Rufino, A. Zheng, S. S. Narayanan, J. Erb, and U. Vaidya, “Modeling quadruped leg dynamics on deformable terrains using data-driven koopman operators,”IFAC-PapersOnLine, vol. 55, no. 37, pp. 420–425, 2022

  21. [21]

    Koopman operator based linear model predictive control for quadruped trotting,

    C.-M. Yang and P. A. Bhounsule, “Koopman operator based linear model predictive control for quadruped trotting,” in2025 IEEE International Conference on Robotics and Automation (ICRA), pp. 12359–12364, IEEE, 2025

  22. [22]

    Adaptive koopman embedding for robust control of nonlinear dynamical systems,

    R. Singh, C. K. Sah, and J. Keshavan, “Adaptive koopman embedding for robust control of nonlinear dynamical systems,”The International Journal of Robotics Research, vol. 44, no. 13, pp. 2235–2261, 2025

  23. [23]

    Model predictive control of nonlinear dynamics using online adaptive koopman operators,

    D. Uchida and K. Duraisamy, “Model predictive control of nonlinear dynamics using online adaptive koopman operators,”arXiv preprint arXiv:2412.02972, 2024

  24. [24]

    Continual learning and lifting of koopman dynamics for linear control of legged robots,

    F. Li, A. Abuduweili, Y . Sun, R. Chen, W. Zhao, and C. Liu, “Continual learning and lifting of koopman dynamics for linear control of legged robots,”arXiv preprint arXiv:2411.14321, 2024

  25. [25]

    Metakoopman: Bayesian meta- learning of koopman operators for modeling structured dynamics under distribution shifts,

    M. Selim, S. Bhat, and K. H. Johansson, “Metakoopman: Bayesian meta- learning of koopman operators for modeling structured dynamics under distribution shifts,” inThe Thirty-ninth Annual Conference on Neural Information Processing Systems

  26. [26]

    Event-based adaptive koopman framework for optic flow-guided landing on moving platforms,

    B. Banday, C. K. Sah, and J. Keshavan, “Event-based adaptive koopman framework for optic flow-guided landing on moving platforms,”arXiv preprint arXiv:2501.16868, 2025

  27. [27]

    Active learning of dynamics for data-driven control using koopman operators,

    I. Abraham and T. D. Murphey, “Active learning of dynamics for data-driven control using koopman operators,”IEEE Transactions on Robotics, vol. 35, no. 5, pp. 1071–1083, 2019

  28. [28]

    Mauroy, Y

    A. Mauroy, Y . Susuki, and I. Mezic,Koopman operator in systems and control, vol. 484. Springer, 2020

  29. [29]

    Gcontrol-aware learning of koopman embedding models a koopman spectral approach,

    D. Goswami and D. A. Paley, “Gcontrol-aware learning of koopman embedding models a koopman spectral approach,” in2017 IEEE 56th Annual Conference on Decision and Control (CDC), pp. 6107–6112, IEEE, 2017

  30. [30]

    V ovk, A

    V . V ovk, A. Gammerman, and G. Shafer,Algorithmic Learning in a Random World. New York, NY: Springer, 2005

  31. [31]

    acados – a modular open-source framework for fast embedded optimal control,

    R. Verschueren, G. Frison, D. Kouzoupis, J. Frey, N. van Duijkeren, A. Zanelli, B. Novoselnik, T. Albin, R. Quirynen, and M. Diehl, “acados – a modular open-source framework for fast embedded optimal control,” Mathematical Programming Computation, 2021