Recognition: unknown
A New Adaptive Deep Learning based Reduced Order Model for Hybrid-Type Parabolic PDEs: Rigorous Error Analysis and Applications
Pith reviewed 2026-05-08 10:38 UTC · model grok-4.3
The pith
Two adaptive deep-learning reduced-order models based on deep orthogonal decomposition provide rigorous error bounds for hybrid parabolic PDEs and outperform traditional POD methods under identifiable conditions.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
We extend the deep orthogonal decomposition method to parabolic PDEs by introducing the DOD-DL-ROM and DOD+DFNN approaches, generalize data-driven POD error arguments to this setting to link online performance with the regularity of the optimal map, and derive conditions on problem size and error tolerance under which DOD-based reduced order models outperform POD-based ones for problems with split parameter dependence.
What carries the argument
The Deep Orthogonal Decomposition (DOD), which separates the solution into components with different parameter sensitivities using deep learning for the low-dimensional dynamics.
Load-bearing premise
The parameter dependence of the parabolic PDE can be split into two parts with distinctly different Kolmogorov N-width decay rates, and this split must be known or identifiable in advance for the methods to remain stable and the error bounds to apply.
What would settle it
Numerical verification on the catalyst filter benchmark problem showing that the approximation error violates the derived a priori or a posteriori bounds, or that the DOD methods fail to outperform POD-ROMs in regimes where the theory predicts superiority.
read the original abstract
This contribution proposes novel data-driven surrogate modeling approaches for parameterized parabolic PDEs, where the parameter dependence can be split into two parts with different decay behavior of the Kolmogorov $N$-width. Such problems naturally arise in many industrial flow processes with dominant advection or traveling fronts in the solution trajectories. To tackle this challenge, we extend the Deep Orthogonal Decomposition (DOD) method, recently introduced for related stationary problems, to the time-dependent setting. We introduce and rigorously analyze two DOD based approaches: Our approach is based on two novel adaptive deep learning-based surrogate models: The DOD-DL-ROM method which is a Reduced Order Model (ROM) that leverages the adaptive nature of DOD, and the DOD+DFNN method, which combines DOD with a generic Deep Feed-Forward Neural Network (DFNN). On the theory side, we generalize data-driven POD-based ROM arguments to the DOD setting, establishing a quantitative link between online performance and the regularity of an associated optimal map. Furthermore, we identify specific problem size and error tolerance requirements for DOD-based ROMs to outperform POD-based ROMs in hybrid-type problem classes, which is crucial for efficient computation. The significance of this work lies in its potential to accelerate the solution of complex PDEs, enabling faster design and optimization of industrial processes. The proposed approaches are demonstrated on a catalyst filter benchmark problem, showcasing their effectiveness and comparing favorably to traditional POD-based methods.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The manuscript proposes two novel adaptive deep learning-based reduced order models (DOD-DL-ROM and DOD+DFNN) for parameterized parabolic PDEs whose parameter dependence splits into two parts exhibiting different Kolmogorov N-width decay rates. It extends the Deep Orthogonal Decomposition (DOD) framework to the time-dependent setting, generalizes data-driven POD-based ROM arguments to DOD, establishes a quantitative link between online performance and the regularity of an associated optimal map, derives conditions on problem size and error tolerance under which DOD-based ROMs outperform POD-based ones, and demonstrates the methods on a catalyst filter benchmark problem.
Significance. If the central claims hold, the work provides a theoretically grounded approach to efficient surrogate modeling for advection- or front-dominated parabolic problems arising in industrial flows. The rigorous generalization of POD arguments, the explicit outperformance thresholds, and the benchmark comparison constitute clear strengths. The paper supplies both error analysis and reproducible numerical demonstration on a standard benchmark.
major comments (3)
- [Introduction and §2] Introduction and §2 (model assumptions): The central claim that the parameter dependence admits an a priori known split into two components with distinctly different N-width decay rates, and that this split remains stable under the time evolution of the parabolic PDE, is load-bearing for all subsequent error bounds and outperformance statements. The manuscript must either supply an explicit identification procedure that does not introduce additional approximation error or prove that the split is preserved by the evolution operator; otherwise the generalization of POD arguments to the adaptive DOD setting and the derived quantitative performance link may fail to hold.
- [Theory section] Theory section (generalization of data-driven POD arguments, main error theorem): The quantitative link between online performance and regularity of the optimal map is stated to follow from the DOD extension, yet the proof sketch appears to treat the N-width split as exact. If the split is only approximate, the stability of the adaptive DOD extension and the resulting a-priori error bounds require an additional perturbation term; the current derivation does not indicate how this term is controlled.
- [§5] §5 (numerical experiments and outperformance thresholds): The identification of specific problem-size and error-tolerance requirements for DOD-based ROMs to outperform POD-based ROMs is presented as a contribution, but the reported thresholds appear to be read off from a single benchmark run. The manuscript should state whether these thresholds were predicted by the theory before the experiments or fitted post hoc, and should include a sensitivity study with respect to small perturbations of the assumed split.
minor comments (2)
- [§3 and figures] Notation for the optimal map and the two DOD-based methods is introduced in the abstract and §3 but is not uniformly used in the figure legends; a short nomenclature table would improve readability.
- [Method description] The DFNN architecture details (depth, width, activation) for the DOD+DFNN variant are described only briefly; adding a table with hyper-parameter choices would aid reproducibility.
Simulated Author's Rebuttal
We thank the referee for the thorough review and constructive comments, which have helped clarify several important aspects of the manuscript. We address each major comment below and have revised the relevant sections accordingly.
read point-by-point responses
-
Referee: [Introduction and §2] Introduction and §2 (model assumptions): The central claim that the parameter dependence admits an a priori known split into two components with distinctly different N-width decay rates, and that this split remains stable under the time evolution of the parabolic PDE, is load-bearing for all subsequent error bounds and outperformance statements. The manuscript must either supply an explicit identification procedure that does not introduce additional approximation error or prove that the split is preserved by the evolution operator; otherwise the generalization of POD arguments to the adaptive DOD setting and the derived quantitative performance link may fail to hold.
Authors: We agree that the a priori split is central. In the manuscript the split is part of the definition of the hybrid-type problem class and is identified directly from the PDE coefficients without introducing approximation error (e.g., parameters multiplying the advective term versus those multiplying the diffusive term). We have added an explicit identification procedure in the revised §2 together with a short lemma showing preservation under the parabolic evolution operator, using the fact that the linear evolution semigroup acts separately on the two parameter-dependent subspaces and that the DOD orthogonality is preserved by the time-stepping scheme. revision: yes
-
Referee: [Theory section] Theory section (generalization of data-driven POD arguments, main error theorem): The quantitative link between online performance and regularity of the optimal map is stated to follow from the DOD extension, yet the proof sketch appears to treat the N-width split as exact. If the split is only approximate, the stability of the adaptive DOD extension and the resulting a-priori error bounds require an additional perturbation term; the current derivation does not indicate how this term is controlled.
Authors: The main theorem is stated for the exact-split case that defines the hybrid-type class. To cover practical situations in which the split is only approximate, we have extended the proof in the revised theory section by inserting a perturbation term whose magnitude is bounded by the product of the split-approximation error and the Lipschitz constant of the evolution operator. The resulting a-priori bound retains the same dependence on the regularity of the optimal map, with an additive constant that can be made arbitrarily small by improving the split identification. revision: yes
-
Referee: [§5] §5 (numerical experiments and outperformance thresholds): The identification of specific problem-size and error-tolerance requirements for DOD-based ROMs to outperform POD-based ROMs is presented as a contribution, but the reported thresholds appear to be read off from a single benchmark run. The manuscript should state whether these thresholds were predicted by the theory before the experiments or fitted post hoc, and should include a sensitivity study with respect to small perturbations of the assumed split.
Authors: The thresholds follow directly from the conditions in Theorem 4.3 (problem dimension versus target tolerance) and were therefore known before the numerical experiments were performed; the catalyst-filter run serves only as confirmation. In the revised §5 we now state this explicitly and add a sensitivity study in which the assumed split is perturbed by 5–10 % mis-assignment of parameters. The results show that the predicted outperformance region remains valid, with only a modest increase in the observed error that stays within the theoretical bound. revision: yes
Circularity Check
Minor self-citation to prior DOD work; central error analysis and outperformance claims rest on external a priori assumptions
full rationale
The derivation extends the DOD method (cited as recently introduced for stationary problems) to time-dependent hybrid parabolic PDEs and generalizes POD-based ROM error arguments to establish a quantitative link between online performance and optimal map regularity. This generalization and the identification of problem-size/error-tolerance requirements for outperformance over POD rely on the stated assumption that the parameter dependence splits into two parts with distinctly different Kolmogorov N-width decay rates, and that this split is known or identifiable a priori. No self-definitional reductions, fitted inputs renamed as predictions, or load-bearing self-citation chains appear in the abstract or described claims; the error bounds and comparisons are presented as holding under those external regularity and split assumptions rather than being forced by internal construction.
Axiom & Free-Parameter Ledger
Reference graph
Works this paper leans on
-
[1]
Proceedings of the Conference Algoritmy, 1–12 (2016)
Ohlberger, M., Rave, S.: Reduced basis methods: Success, limitations and future challenges. Proceedings of the Conference Algoritmy, 1–12 (2016)
2016
-
[2]
(eds.): Model Reduction and Approximation
Benner, P., Cohen, A., Ohlberger, M., Willcox, K. (eds.): Model Reduction and Approximation. Computational Science & Engineering, vol. 15, p. 412. SIAM, Philadelphia, PA, USA (2017). https://doi.org/10.1137/1.9781611974829 . Theory and algorithms
-
[3]
(eds.): Model Reduction of Parametrized Systems
Benner, P., Ohlberger, M., Patera, A., Rozza, G., Urban, K. (eds.): Model Reduction of Parametrized Systems. MS&A. Modeling, Simulation and Applica- tions, vol. 17, p. 504. Springer, Cham, Germany (2017). https://doi.org/10.1007/ 978-3-319-58786-8 . Selected papers from the 3rd MoRePaS Conference held at the International School for Advanced Studies (SISS...
2017
-
[4]
In: Yaglom, A.M., Tatarsky, V.I
Lumley, J.L.: The structure of inhomogeneous turbulent flows. In: Yaglom, A.M., Tatarsky, V.I. (eds.) Atmospheric Turbulence and Radio Wave Propagation, pp. 166–178. Nauka, Moscow (1967)
1967
-
[5]
part i: Coherent structures
Sirovich, L.: Turbulence and the dynamics of coherent structures. part i: Coherent structures. Journal of Fluid Mechanics175, 419–427 (1987) https://doi.org/10. 1017/S0022112087000199
1987
-
[6]
Cambridge University Press, Cambridge (1996)
Holmes, P., Lumley, J.L., Berkooz, G.: Turbulence, Coherent Structures, Dynam- ical Systems and Symmetry. Cambridge University Press, Cambridge (1996)
1996
-
[7]
SIAM Journal on Numerical Analysis 40(2), 492–515 (2002)
Kunisch, K., Volkwein, S.: Galerkin proper orthogonal decomposition methods for a general equation in fluid dynamics. SIAM Journal on Numerical Analysis 40(2), 492–515 (2001) https://doi.org/10.1137/S0036142900382612
-
[8]
Haasdonk, B., Ohlberger, M.: Reduced basis method for finite volume approx- imations of parametrized linear evolution equations. Mathematical Modelling and Numerical Analysis42(2), 277–302 (2008) https://doi.org/10.1051/m2an: 2008001
-
[9]
Barrault, M., Maday, Y., Nguyen, N.C., Patera, A.T.: An ’empirical interpolation’ method: application to efficient reduced-basis discretization of partial differential equations. Comptes Rendus Mathematique339(9), 667–672 (2004) https://doi. org/10.1016/j.crma.2004.08.006
-
[10]
Chaturantabut, S., Sorensen, D.C.: Nonlinear model reduction via discrete empiri- cal interpolation. SIAM Journal on Scientific Computing32(5), 2737–2764 (2010) https://doi.org/10.1137/090766498
-
[11]
Drohmann, M., Haasdonk, B., Ohlberger, M.: Reduced basis approximation for 32 nonlinear parametrized evolution equations based on empirical operator inter- polation. SIAM J. Sci. Comput.34(2), 937–969 (2012) https://doi.org/10.1137/ 10081157X
2012
-
[12]
Carlberg, K., Bou-Mosleh, C., Farhat, C.: Efficient non-linear model reduction via a least-squares petrov-galerkin projection and compressive tensor approximations. International Journal for Numerical Methods in Engineering86, 155–181 (2011) https://doi.org/10.1002/nme.3050
-
[13]
Journal of Computational Physics242, 623–647 (2013) https://doi.org/10.1016/j.jcp.2013.02.028
Carlberg, K., Farhat, C., Cortial, J., Amsallem, D.: The gnat method for nonlinear model reduction: Effective implementation and application to computational fluid dynamics and turbulent flows. Journal of Computational Physics242, 623–647 (2013) https://doi.org/10.1016/j.jcp.2013.02.028
-
[14]
In: Proceedings of the 61st Annual Meeting of the APS Division of Fluid Dynamics (2010)
Schmid, P.J.: Dynamic mode decomposition of numerical and experimental data. In: Proceedings of the 61st Annual Meeting of the APS Division of Fluid Dynamics (2010)
2010
-
[15]
Spectral analysis of nonlinear flows,
Rowley, C.W., Mezi´ c, I., Bagheri, S., Schlatter, P., Henningson, D.S.: Spectral analysis of nonlinear flows. Journal of Fluid Mechanics641, 115–127 (2009) https: //doi.org/10.1017/S0022112009992059
-
[16]
Peherstorfer, B., Willcox, K., Gunzburger, M.: Survey of multifidelity methods in uncertainty propagation, inference, and optimization. SIAM Review60, 550–591 (2018) https://doi.org/10.1137/16M1082469
-
[17]
https://arxiv.org/abs/2001.04001
Fresca, S., Dede, L., Manzoni, A.: A comprehensive deep learning-based approach to reduced order modeling of nonlinear time-dependent parametrized PDEs (2020). https://arxiv.org/abs/2001.04001
-
[18]
Fresca, S., Manzoni, A.: Pod-dl-rom: Enhancing deep learning-based reduced order models for nonlinear parametrized pdes by proper orthogonal decompo- sition. Computer Methods in Applied Mechanics and Engineering388, 114181 (2022) https://doi.org/10.1016/j.cma.2021.114181
-
[19]
https: //arxiv.org/abs/2305.04680
Brivio, S., Fresca, S., Franco, N.R., Manzoni, A.: Error estimates for POD- DL-ROMs: a deep learning framework for reduced order modeling of nonlinear parametrized PDEs enhanced by proper orthogonal decomposition (2023). https: //arxiv.org/abs/2305.04680
-
[20]
Haasdonk, B., Kleikamp, H., Ohlberger, M., Schindler, F., Wenzel, T.: A new cer- tified hierarchical and adaptive RB-ML-ROM surrogate model for parametrized PDEs. SIAM J. Sci. Comput.45(3), 1039–1065 (2023) https://doi.org/10.1137/ 22M1493318
2023
-
[21]
Klein, B., Ohlberger, M.: Multi-fidelity learning of reduced order models for parabolic pde constrained optimization. Adv. Comput. Math.52(19), 1–36 (2026) 33 https://doi.org/10.1007/s10444-026-10296-6
-
[22]
Lecture Notes in Computer Science13127 LNCS, 378–386 (2022) https://doi.org/10.1007/ 978-3-030-97549-4 43
Gavrilenko, P., Haasdonk, B., Iliev, O., Ohlberger, M., Schindler, F., Toktal- iev, P., Wenzel, T., Youssef, M.: A full order, reduced order and machine learning model pipeline for efficient prediction of reactive flows. Lecture Notes in Computer Science13127 LNCS, 378–386 (2022) https://doi.org/10.1007/ 978-3-030-97549-4 43
2022
-
[23]
Fokina, D., Toktaliev, P., Herkert, R., Wenzel, T., Haasdonk, B., Iliev, O.: Machine learning methods based prediction of breakthrough curves in reactive porous media from peclet and damk¨ ohler numbers. Studies in Computational Intelligence1219 SCI, 67–84 (2025) https://doi.org/10.1007/978-3-031-96311-7 7
-
[24]
Lecture Notes in Computer Science13952 LNCS, 91–98 (2024) https://doi.org/10.1007/ 978-3-031-56208-2 8
Fokina, D., Grigoriev, V.V., Iliev, O., Oseledets, I.: Machine learning algo- rithms for parameter identification for reactive flow in porous media. Lecture Notes in Computer Science13952 LNCS, 91–98 (2024) https://doi.org/10.1007/ 978-3-031-56208-2 8
2024
-
[25]
Chemical Engineering Journal 519, 163598 (2025) https://doi.org/10.1016/j.cej.2025.163598
Biermann, F., Uglietti, R., D¨ oppel, F.A., Kircher, T., Bracconi, M., Maestri, M., Votsmeier, M.: Enabling micro-kinetics based simulation of industrial packed- bed reactors by physics-enhanced neural networks. Chemical Engineering Journal 519, 163598 (2025) https://doi.org/10.1016/j.cej.2025.163598
-
[26]
Preprint, arXiv:2404.18841 [math.NA] (2024)
Franco, N.R., Manzoni, A., Zunino, P., Hesthaven, J.S.: Deep orthogonal decom- position: a continuously adaptive data-driven approach to model order reduction. Preprint, arXiv:2404.18841 [math.NA] (2024). arXiv:2404.18841 (2024). https: //arxiv.org/abs/2404.18841
-
[27]
Error bounds for approximations with deep
Yarotsky, D.: Error bounds for approximations with deep relu networks. Neural Networks94, 103–114 (2017) https://doi.org/10.1016/j.neunet.2017.07.002
-
[28]
https://arxiv.org/abs/1902
G¨ uhring, I., Kutyniok, G., Petersen, P.: Error bounds for approximations with deep ReLU neural networks inW s,p norms (2019). https://arxiv.org/abs/1902. 07896
2019
-
[29]
Neural Networks161, 129– 141 (2023) https://doi.org/10.1016/j.neunet.2023.01.029
Franco, N.R., Fresca, S., Manzoni, A., Zunino, P.: Approximation bounds for convolutional neural networks in operator learning. Neural Networks161, 129– 141 (2023) https://doi.org/10.1016/j.neunet.2023.01.029
-
[30]
The SMAI Journal of computational mathematics7, 121–157 (2021) https://doi.org/10.5802/smai-jcm.74
Bhattacharya, K., Hosseini, B., Kovachki, N.B., Stuart, A.M.: Model Reduction And Neural Networks For Parametric PDEs. The SMAI Journal of computational mathematics7, 121–157 (2021) https://doi.org/10.5802/smai-jcm.74
-
[31]
https://arxiv.org/abs/2103
Marwah, T., Lipton, Z.C., Risteski, A.: Parametric Complexity Bounds for Approximating PDEs with Neural Networks (2021). https://arxiv.org/abs/2103. 02138 34
2021
-
[32]
Franco, N.R., Manzoni, A., Zunino, P.: A deep learning approach to reduced order modelling of parameter dependent partial differential equations92, 483–524 (2023) https://doi.org/10.1090/mcom/3781
-
[33]
Neural Networks4(2), 251–257 (1991) https://doi.org/10.1016/0893-6080(91) 90009-T
Hornik, K.: Approximation capabilities of multilayer feedforward networks. Neural Networks4(2), 251–257 (1991) https://doi.org/10.1016/0893-6080(91) 90009-T
-
[34]
Golub, G.H., Van Loan, C.F.: Matrix Computations. Johns Hopkins University Press, ??? (2013). https://doi.org/10.56021/9781421407944
-
[35]
Newey, W.K., McFadden, D.: Chapter 36 large sample estimation and hypothesis testing. Handbook of Econometrics, vol. 4, pp. 2111–2245. Elsevier (1994). https: //doi.org/10.1016/S1573-4412(05)80005-4
-
[36]
https://arxiv.org/abs/2004.12131
Geist, M., Petersen, P., Raslan, M., Schneider, R., Kutyniok, G.: Numerical Solu- tion of the Parametric Diffusion Equation by Deep Neural Networks (2020). https://arxiv.org/abs/2004.12131
-
[37]
In: Interna- tional Conference on Learning Representations (ICLR) (2019)
Loshchilov, I., Hutter, F.: Decoupled weight decay regularization. In: Interna- tional Conference on Learning Representations (ICLR) (2019)
2019
-
[38]
https://arxiv.org/abs/ 2202.06373
Al-Kababji, A., Bensaali, F., Dakua, S.P.: Scheduling Techniques for Liver Seg- mentation: ReduceLRonPlateau Vs OneCycleLR (2022). https://arxiv.org/abs/ 2202.06373
-
[39]
https: //arxiv.org/abs/1909.13672
Bastian, P., Blatt, M., Dedner, A., Dreier, N.-A., Engwer, C., Fritze, R., Gr¨ aser, C., Gr¨ uninger, C., Kempf, D., Kl¨ ofkorn, R., Ohlberger, M., Sander, O.: The DUNE Framework: Basic Concepts and Recent Developments (2020). https: //arxiv.org/abs/1909.13672
-
[40]
Kybernetika 2(2010)
Bastian, P., Heimann, F., Marnach, S.: Generic implementation of finite element methods in the distributed and unified numerics environment (dune). Kybernetika 2(2010)
2010
-
[41]
SIAM Journal on Scientific Computing38(5), 194–216 (2016) https://doi.org/10.1137/15M1026614
Milk, R., Rave, S., Saak, J.: pymor – generic algorithms and interfaces for model order reduction. SIAM Journal on Scientific Computing38(5), 194–216 (2016) https://doi.org/10.1137/15M1026614
-
[42]
Harris, C.R., Millman, K.J., Walt, S.J., Gommers, R., Virtanen, P., Cournapeau, D., Wieser, E., Taylor, J., Berg, S., Smith, N.J., Kern, R., Picus, M., Hoyer, S., Kerkwijk, M.H., Brett, M., Haldane, A., R´ ıo, J.F., Wiebe, M., Peterson, P., G´ erard-Marchant, P., Sheppard, K., Reddy, T., Weckesser, W., Abbasi, H., Gohlke, C., Oliphant, T.E.: Array program...
-
[43]
PyTorch: An Imperative Style, High-Performance Deep Learning Library
Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., 35 Lin, Z., Gimelshein, N., Antiga, L., Desmaison, A., K¨ opf, A., Yang, E., DeVito, Z., Raison, M., Tejani, A., Chilamkurthy, S., Steiner, B., Fang, L., Bai, J., Chin- tala, S.: PyTorch: An Imperative Style, High-Performance Deep Learning Library (2019). https://arxiv.org...
work page internal anchor Pith review arXiv 2019
-
[44]
Springer, Cham (2015)
Quarteroni, A., Manzoni, A., Negri, F.: Reduced Basis Methods for Par- tial Differential Equations. Springer, Cham (2015). https://doi.org/10.1007/ 978-3-319-15431-2
2015
-
[45]
Kuratowski, K., Ryll-Nardzewski, C.: A general theorem on selectors. Bull. Acad. Polon. Sci. S´ er. Sci. Math. Astronom. Phys.13, 397–403 (1965) Appendix A Reduced Basis Framework For this subsection we defineu(µ j, tk)∈R Nh for (µ j)Ns j=1 =:P ⊂Θ and (t k)Nt k=1 =: T ⊂[0, T] to be some vectors. They will later represent the solutions to a parametric PDE ...
1965
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.