pith. machine review for the scientific record. sign in

arxiv: 2605.01305 · v1 · submitted 2026-05-02 · 🧮 math.NA · cs.NA

Recognition: unknown

Alikhanov-XfPINNs: Adaptive Physics-Informed Learning for Nonlinear Fractional PDEs on Nonuniform Meshes

Himanshu Kumar Dwivedi, Matthias Ehrhardt, Rajeev

Authors on Pith no claims yet

Pith reviewed 2026-05-09 18:33 UTC · model grok-4.3

classification 🧮 math.NA cs.NA
keywords nonlinear fractional PDEsphysics-informed neural networksAlikhanov discretizationnonuniform time gridsinitial singularityforward and inverse problemsvariational loss function
0
0 comments X

The pith

The Alikhanov-XfPINNs architecture embeds high-order time discretization on nonuniform meshes into neural network training to solve nonlinear fractional PDEs accurately despite initial singularities.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

This paper develops XfPINNs to solve nonlinear fractional partial differential equations that suffer from initial singularities and high computational costs due to nonlocal memory terms. It combines an accelerated Alikhanov discretization scheme on nonuniform time grids with the physics-informed neural networks framework, embedding the variational loss directly into training. Adaptive activation functions speed convergence while a mix of hard and soft constraints handles initial and boundary conditions. The result supports both forward solution approximation and inverse parameter estimation, with an auxiliary time-marching setup that isolates temporal discretization effects from training artifacts.

Core claim

The authors establish that the XfPINNs framework, built by extending fractional PINNs with the Alikhanov variational loss on adaptive nonuniform meshes, successfully approximates solutions to general nonlinear fractional PDEs. It supports both forward simulations and inverse parameter identification, and numerical tests confirm that accuracy tracks the temporal mesh refinement while training remains stable and efficient.

What carries the argument

The Alikhanov-extended fractional PINNs (XfPINNs) architecture that constructs a variational loss from the high-order Alikhanov time-stepping scheme applied on nonuniform grids and trains the network to minimize it while enforcing initial and boundary conditions through hard and soft penalties.

If this is right

  • Solutions to nonlinear fPDEs can be reconstructed from data even when the exact form is unknown.
  • Parameter estimation in inverse problems becomes feasible with the same trained architecture.
  • Computational efficiency improves substantially through CPU time reductions on nonuniform meshes compared to uniform ones.
  • Temporal convergence rates can be isolated and verified independently of neural network optimization errors.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • This approach might generalize to other singular fractional models in physics and engineering if the nonuniform grid strategy adapts well.
  • By auditing discretization separately, it provides a template for testing other hybrid numerical-ML methods for PDEs.
  • Future work could explore making the mesh adaptation fully learnable within the network to further optimize accuracy.

Load-bearing premise

Embedding the Alikhanov variational loss into the neural network training along with hard and soft constraints for conditions ensures that solution accuracy is governed mainly by the choice of temporal discretization rather than by issues in the optimization process or data sampling on nonuniform meshes.

What would settle it

An experiment showing that refining the nonuniform time grid does not reduce the overall error below a certain level set by training tolerances, or that increasing network capacity fails to improve results when the mesh is fixed, would indicate the claim does not hold.

Figures

Figures reproduced from arXiv: 2605.01305 by Himanshu Kumar Dwivedi, Matthias Ehrhardt, Rajeev.

Figure 1
Figure 1. Figure 1: Standard activation functions modified adaptively by varying the hyperparam view at source ↗
Figure 2
Figure 2. Figure 2: Schematic diagram of the XfPINN architecture. view at source ↗
Figure 3
Figure 3. Figure 3: Comparison of loss evolution and Alikhanov–XfPINNs approximation results for view at source ↗
Figure 4
Figure 4. Figure 4: Evolution of the training loss and na across iterations with α = 0.5 for adaptive activation functions with various n : (a) loss trajectory; (b) na. 21 view at source ↗
Figure 5
Figure 5. Figure 5: Comparison of 2D-NTFSDe with smooth exact solution at view at source ↗
Figure 6
Figure 6. Figure 6: Comparison of exact solution v, Alikhanov–XfPINNs prediction v˜ and absolute error |v − v˜| for time fractional generalized Burgers’ problem. 24 view at source ↗
Figure 7
Figure 7. Figure 7: Alikhanov–XfPINN predictions for α = 0.3, 0.6, 0.9 vs. exact integer-order FN solutions: (a) 1D-TFFN; (b) 2D-TFFN. For the inverse problem, the training dataset is extended by incorporating 30 randomly sampled terminal time solutions into the original forward dataset view at source ↗
Figure 8
Figure 8. Figure 8: Evolution of the estimated parameters α, λ1, and λ2 during training of the Alikhanov–XfPINNs for the 2D-TFRD equation on nonuniform meshes with varying grad￾ing parameter γ view at source ↗
Figure 9
Figure 9. Figure 9: Evolution of the estimated parameters α, ε, and Γ during training of the Alikhanov–XfPINNs for the 2D-TFAC equation on nonuniform meshes with varying grad￾ing parameter γ. 4 Conclusion This work introduces a data-driven approach to solving nonlinear fractional partial differential equa￾tions (NfPDEs) that integrates an accelerated Alikhanov discretization on nonuniform time meshes with physics-informed neu… view at source ↗
read the original abstract

To address the initial singularity inherent in solutions to fractional partial differential equations (fPDEs), we propose an accelerated Alikhanov discretization formulation implemented on nonuniform time grids. Based on the physics-informed neural networks (PINNs) framework, we introduce an Alikhanov-extended fractional PINNs (XfPINNs) architecture that combines high-order temporal discretization and deep learning. The nonlocal memory term in fPDEs leads to high computational cost, while the weak singularity near $t\to 0^+$ can deteriorate accuracy on uniform meshes. To separate temporal discretization effects from optimization and sampling errors, we further develop an auxiliary time-marching configuration that enables auditable temporal-convergence studies under controlled training tolerances. This architecture can solve general nonlinear fPDEs. The XfPINNs approach is designed for forward and inverse problems, allowing for data-driven solution reconstruction and parameter estimation. First, the neural network approximates the solution of nonlinear fPDEs; then, an adaptive activation function accelerates convergence and enhances training efficiency. The optimization framework embeds a variational loss function constructed from the Alikhanov scheme, where the initial and boundary conditions are imposed using a combination of hard and soft constraints. Numerical experiments, including cases with known and unknown exact solutions which demonstrate the robustness, computational efficiency, and significant CPU time savings of the Alikhanov-XfPINNs method.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 1 minor

Summary. The manuscript proposes Alikhanov-XfPINNs, which embeds an accelerated Alikhanov discretization on nonuniform time grids into the PINNs framework to solve nonlinear fractional PDEs while addressing the initial singularity at t=0+. An auxiliary time-marching configuration is introduced to separate temporal discretization effects from optimization and sampling errors, enabling controlled convergence studies. The architecture uses adaptive activation functions, a variational loss derived from the Alikhanov scheme, and a mix of hard and soft constraints for IC/BCs. It is applied to both forward and inverse problems, with numerical experiments claimed to demonstrate robustness, computational efficiency, and substantial CPU time savings.

Significance. If the error-separation claim holds and the attained training residuals remain below the O(τ^{2-α}) discretization error on graded meshes, the work would provide a practical hybrid method for nonlocal fPDEs that combines high-order temporal accuracy with the flexibility of neural networks for data-driven forward and inverse tasks, potentially offering measurable CPU savings over purely mesh-based or standard PINN approaches.

major comments (2)
  1. [auxiliary time-marching configuration] The central claim that accuracy is limited primarily by the temporal discretization rather than by optimization or sampling artifacts rests on the auxiliary time-marching configuration. No explicit verification is supplied showing that the minimized loss is smaller than the expected discretization error independently of mesh grading, network capacity, or the strength of the singularity at t=0+. This verification is load-bearing for the robustness and auditable-convergence assertions.
  2. [numerical experiments] Numerical experiments section: while the abstract states that experiments demonstrate robustness, efficiency, and CPU savings, the visible description provides no quantitative error tables, observed convergence rates, or direct baseline comparisons (e.g., against standard PINNs or graded-mesh finite-difference schemes). Without these, the efficiency claims cannot be assessed against the asserted O(τ^{2-α}) behavior.
minor comments (1)
  1. [methods] The notation for the variational loss and the precise definition of the adaptive activation function should be stated explicitly with equation numbers for reproducibility.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the constructive and detailed comments on our manuscript. We address each major comment below and will revise the manuscript to strengthen the presentation and supporting evidence.

read point-by-point responses
  1. Referee: [auxiliary time-marching configuration] The central claim that accuracy is limited primarily by the temporal discretization rather than by optimization or sampling artifacts rests on the auxiliary time-marching configuration. No explicit verification is supplied showing that the minimized loss is smaller than the expected discretization error independently of mesh grading, network capacity, or the strength of the singularity at t=0+. This verification is load-bearing for the robustness and auditable-convergence assertions.

    Authors: We agree that explicit verification is essential to support the central claim regarding the auxiliary time-marching configuration. The manuscript introduces this configuration specifically to enable controlled temporal-convergence studies by separating discretization effects from optimization and sampling errors. However, we acknowledge that additional checks demonstrating the minimized loss remains below the O(τ^{2-α}) discretization error—independent of mesh grading, network capacity, and singularity strength—are not sufficiently detailed. In the revised manuscript, we will add targeted numerical verifications, including comparative results across these factors, to make the auditable-convergence assertions more robust. revision: yes

  2. Referee: [numerical experiments] Numerical experiments section: while the abstract states that experiments demonstrate robustness, efficiency, and CPU savings, the visible description provides no quantitative error tables, observed convergence rates, or direct baseline comparisons (e.g., against standard PINNs or graded-mesh finite-difference schemes). Without these, the efficiency claims cannot be assessed against the asserted O(τ^{2-α}) behavior.

    Authors: We accept this criticism of the numerical experiments section. Although the manuscript describes experiments on forward and inverse problems with known and unknown solutions and asserts robustness, efficiency, and CPU savings, we recognize that quantitative error tables, explicit convergence rates, and direct baseline comparisons are not provided in sufficient detail. In the revision, we will expand the section to include comprehensive error tables, plots of observed convergence rates verifying the O(τ^{2-α}) behavior on graded meshes, and comparisons against standard PINNs and graded-mesh finite-difference schemes to allow proper assessment of the efficiency claims. revision: yes

Circularity Check

0 steps flagged

No circularity: derivation relies on external Alikhanov scheme and standard PINN loss without self-referential reduction

full rationale

The paper constructs XfPINNs by embedding the known Alikhanov variational discretization (a prior external scheme) into a standard physics-informed neural network loss, augmented with adaptive activations, hard/soft constraints, and an auxiliary time-marching setup for controlled convergence studies. No equation or claim reduces the asserted accuracy, efficiency, or error control to a fitted parameter or quantity defined by the method itself; the auxiliary configuration is an independent design choice to isolate discretization order from optimization artifacts, and numerical experiments provide external validation rather than tautological confirmation. The architecture therefore remains self-contained against external benchmarks.

Axiom & Free-Parameter Ledger

0 free parameters · 2 axioms · 0 invented entities

The central claim rests on the assumption that the Alikhanov scheme remains stable and accurate when its variational form is used as a neural-network loss on nonuniform grids, plus standard neural-network approximation capabilities for the solution space.

axioms (2)
  • domain assumption The Alikhanov discretization provides a consistent high-order approximation to the fractional derivative on nonuniform meshes with controlled truncation error near t=0.
    Invoked when constructing the variational loss function from the Alikhanov scheme.
  • domain assumption A neural network with adaptive activation can approximate the solution of the nonlinear fPDE sufficiently well that the embedded discretization error dominates the total error.
    Underlies the claim that temporal discretization effects can be isolated from optimization and sampling errors.

pith-pipeline@v0.9.0 · 5561 in / 1499 out tokens · 20582 ms · 2026-05-09T18:33:37.663441+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

43 extracted references · 1 canonical work pages

  1. [1]

    Krizhevsky, I

    A. Krizhevsky, I. Sutskever, G. E. Hinton, ImageNet classification with deep convolutional neural networks, Adv. Neural Process. Syst. (2012) 1097-1105

  2. [2]

    B. M. Lake, R. Salakhutdinov, J. B. Tenenbaum, Human-level concept learning through prob- abilistic program induction, Science 350 (2015) 1332-1338

  3. [3]

    Raissi, P

    M. Raissi, P. Perdikaris, G. E. Karniadakis, Physics-informed neural networks: A deep learn- ing framework for solving forward and inverse problem involving nonlinear partial differential equations, J. Comput. Phys. 378 (2019) 686-707. 29

  4. [4]

    G. E. Karniadakis, I. G. Kevrekidis, L. Lu, P. Perdikaris, S. Wang, L. Yang, Physics-informed machine learning, Nat. Rev. Phys. 3(6) (2021) 422-440

  5. [5]

    X. Jin, S. Cai, H. Li, G. E. Karniadakis, NSFnets (Navier-Stokes flow nets): Physics-informed neural networks for the incompressible Navier-Stokes equations, J. Comput. Phys. 426 (2021) 109951

  6. [6]

    Raissi, Z

    M. Raissi, Z. Wang, M. S. Triantafyllou, G. E. Karniadakis, Deep learning of vortex-induced vibrations, J. Fluid Mech. 861 (2019) 119-137

  7. [7]

    D. Liu, Y. Wang, Multi-fidelity physics-constrained neural network and its application in ma- terial modeling, J. Mech. Des. 141(12) (2019) 121403

  8. [8]

    Kissas, Y

    G. Kissas, Y. Yang, E. Hwuang, W. R. Witschey, J. A. Detre, P. Perdikaris, Machine learning in cardiovascular flows modeling: predicting arterial blood pressure from non-invasive 4D flow MRI data using physics-informed neural networks, Comput. Methods Appl. Mech. Eng. 358 (2020) 112623

  9. [9]

    Sirignano, K

    J. Sirignano, K. Spiliopoulos, DGM: A deep learning algorithm for solving partial differential equations, J. Comput. Phys. 375 (2018) 1339-1364

  10. [10]

    L. Lu, X. Meng, Z. Mao, G. E. Karniadakis, DeepXDE: a deep learning library for solving differential equations, SIAM Rev. 63(1) (2021) 208-228

  11. [11]

    S. Wang, S. Sankaran, P. Perdikaris, Respecting causality for training physics-informed neural networks, Comput. Methods Appl. Mech. Eng Phys. 421 (2024) 116813

  12. [12]

    Basdevant, M

    C. Basdevant, M. Deville, P. Haldenwang, J. Lacorix, J. Ouazzani, R. Peyret, P. Orlandi, A. Patera, Spectral and finite difference solutions of the Burgers’ equation, Comput. Fluids. 14 (1986) 23-41

  13. [13]

    Z. Sun, G. Gao, Fractional differential equations: finite difference methods, De Gruyter, Berlin (2020)

  14. [14]

    B. Jin, R. Lazarov, Z. Zhou, An analysis of the L1 scheme for the subdiffusion equation with nonsmooth data, IMA J. Numer. Anal. 36 (2016) 197–221

  15. [15]

    Stynes, E

    M. Stynes, E. O´Riordan, J. Gracia, Error analysis of a finite difference method on graded meshes for a time-fractional diffusion equation, SIAM J. Numer. Anal. 55(2) (2017) 1057– 1079

  16. [16]

    McLean, Regularity of solutions to a time-fractional diffusion equation, ANZIAM J

    W. McLean, Regularity of solutions to a time-fractional diffusion equation, ANZIAM J. 52 (2010) 123–138

  17. [17]

    H. Chen, M. Stynes, Error analysis of a second-order method on fitted meshes for a time- fractional diffusion problem, J. Sci. Comput. 79(1) (2019) 624–647

  18. [18]

    Jiang, J

    S. Jiang, J. Zhang, Q. Zhang, Z. Zhang, Fast evaluation of the Caputo fractional derivative and its applications to fractional differential equations, Commun. Comput. Phys. 21(3) (2017). 650-678. 30

  19. [19]

    H. K. Dwivedi, Rajeev, A novel fast second order approach with high-order compact difference scheme and its analysis for the tempered fractional Burgers equation, Math. Comput. Simul. 227 (2025) 168-188

  20. [20]

    H. K. Dwivedi, Rajeev, A novel fast tempered algorithm with high-accuracy scheme for 2D tempered fractional reaction-advection-subdiffusion equation, Comput. Math. Appl. 176 (2024) 371-397

  21. [21]

    J. Cao, A. Xiao, W. Bu, Finite difference/finite element method for the tempered time fractional advection-dispersion equation with fast evaluation of Caputo derivatives, J. Sci. Comput. (2020) 83:48

  22. [22]

    Y. Yan, Z. Z. Sun, J. Zhang, Fast evaluation of the Caputo fractional derivative and its ap- plications to fractional diffusion equations: a second order scheme, Commun. Comput. Phys. 22(4) (2017) 1028-1048

  23. [23]

    X. Li, H. L. Liao, L. Zhang, A second-order fast compact scheme with unequal time-steps for subdiffusion problems, Numer. Algor. 86 (2021) 1011–1039

  24. [24]

    Luchko, W

    Y. Luchko, W. Rundell, M. Yamamoto, L. Zuo, Uniqueness and reconstruction of an unknown semilinear term in a time-fractional reaction diffusion equation, Inverse Probl. 29(6) (2013) 065019

  25. [25]

    Kern, Numerical methods for inverse problems, Wiley, Hoboken (2016)

    M. Kern, Numerical methods for inverse problems, Wiley, Hoboken (2016)

  26. [26]

    Lukyanenko, R

    D. Lukyanenko, R. Argun, A. Borzunov, A. Gorbachev, V. Shinkarev, M. Shislenin, A. Yagola, On the features of numerical solution of coefficient inverse problems for nonlinear equations of the reaction-diffusion-advection type with data of various types, Differ. Equ. 59(12) (2023) 1734-1757

  27. [27]

    X. An, Q. Wang, F. Liu, V. V. Anh, I. W. Turner, Parameter estimation for time-fractional Black-Scholes equation with S&P 500 index option, Numer. Algor. 95(1) (2024) 1-30

  28. [28]

    Srati, A

    M. Srati, A. Oulmelk, L. Afraites, A. Hadri, M. A. Zaky, A. Aldraiweesh, A. S. Hendy, An in- verse problem of determining the parameters in diffusion equations by using fractional physics- informed neural networks, Appl. Numer. Math. 208 (2025) 189-213

  29. [29]

    L. Ma, R. Li, F. Zeng, L. Guo, G. E. Karniadakis, Bi-orthogonal fPINN: a physics-informed neural network method for solving time-dependent stochastic fractional PDEs, Commun. Comput. Phys. 34(4) (2023) 1133-1176

  30. [30]

    X. Fang, L. Qiao, F. Zhang, F. Sun, Explore deep network for a class of fractional partial differential equations, Chaos Solitons Fractals 172 (2022) 113528

  31. [31]

    J. Shi, X. Liu, X, Yang, Data-driven solutions and parameter estimation of the high-dimensional time-fractional reaction-diffusion equations using an improved fPINN method, Nonlinear Dyn. 113 (2025) 9577-9604

  32. [32]

    G. Pang, L. Lu, G. E. Karniadakis, fPINNS: fractional physics-informed neural networks, SIAM J. Sci. Comput. 41(4) (2019) 2603-2626 . 31

  33. [33]

    S. Wang, H. Zhang, X. Jiang, Fractional physics-informed neural networks for time-fractional phase field models, Nonlin. Dyn. 110(3) (2022) 2715-2739

  34. [34]

    Z. Ma, J. Hou, W. Zhu, Y. Peng, Y. Li, PMNN: physical model driven neural network for solving time-fractional differential equations, Chaos Solitons Fractals 177 (2023) 114238

  35. [35]

    H. Ren, X. Meng, R. Liu, J. Hou, Y. Yu, A class of improved fractional physics-informed neural networks, Neurocomputing 562 (2023) 126890

  36. [36]

    H. L. Liao, , D. Li, J. Zhang, Sharp error estimate of nonuniform L1 formula for time-fractional reaction-subdiffusion equations, SIAM J. Numer. Anal. 56 (2018) 1112–1133

  37. [37]

    A. A. Alikhanov, A new difference scheme for the time fractional diffusion equation, J. Comput. Phys. 280, (2015) 424-438

  38. [38]

    H. L. Liao, W. Mclean, J. Zhang, A second-order scheme with nonuniform time steps for a linear reaction-subdiffusion problem, Commun. Comput. Phys. 30(2), (2021) 567-601

  39. [39]

    A. D. Jagtap, K. Kawaguchi, G. E. Karniadakis, Adaptive activation functions accelerate convergence in deep and physics-informed neural networks, J. Comput. Phys. 404, (2020) 109136

  40. [40]

    H. Gao, L. Sun, J. Wang, PhyGeoNet: physics-informed geometry-adaptive convolutional neu- ral networks for solving parametrized steady-state PDE on irregular domain, J. Comput. Phys. 428, (2021) 110079

  41. [41]

    J. Shi, X. Liu, X. Yang, Data-driven solutions and parameter estimation of the high-dimensional time-fractional reaction-diffusion equations using an improved fPINN method, Nonlin. Dyn. 113 (2025)9577-9604

  42. [42]

    TensorFlow: Large-Scale Machine Learning on Heterogeneous Distributed Systems

    M. Abadi, A. Agarwal, P. Barham, E. Brevdeo, Z. Chen, C. Citro, G. S. Corrado, A. Davis, J. Dean, M. Devin, et al., Tensorflow: Large-scale machine learning on heterogeneous distributed systems, arXiv:1603.04467 (2016)

  43. [43]

    Barkat, P

    A. Barkat, P. Bianchi, Convergence and dynamical behavior of the ADAM algorithm for non- convex stochastic optimization, SIAM J. Optim. 31(1) (2021) 244-274. 32