pith. machine review for the scientific record. sign in

arxiv: 2605.12368 · v1 · submitted 2026-05-12 · 💻 cs.LG

Recognition: 2 theorem links

· Lean Theorem

MetaColloc: Optimization-Free PDE Solving via Meta-Learned Basis Functions

Zichuan Yang

Pith reviewed 2026-05-13 05:53 UTC · model grok-4.3

classification 💻 cs.LG
keywords PDE solvingmeta-learningbasis functionscollocationoptimization-freeGaussian random fieldsneural networks
0
0 comments X

The pith

MetaColloc solves new PDEs by freezing meta-learned basis functions and performing one linear least-squares solve at test time.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

Standard neural approaches to PDEs require training a fresh network for each new equation, which is slow. MetaColloc separates the learning of basis functions from the solving step. A dual-branch network is meta-trained once on many Gaussian random fields to produce a reusable dictionary of neural basis functions. At test time the network stays frozen and the PDE is turned into a collocation matrix whose solution is found by a single linear least-squares step, or by Newton iteration for nonlinear cases. The method reports state-of-the-art accuracy on smooth and nonlinear problems while cutting test-time computation by orders of magnitude, and a frequency analysis shows where high-frequency operator stability breaks down.

Core claim

A meta-trained dual-branch neural network produces a universal dictionary of basis functions from Gaussian random fields; once frozen, any new PDE is solved by assembling a collocation matrix from those functions and finding the coefficients through one linear least-squares solve, with Newton-Raphson delivering quadratic convergence on nonlinear equations.

What carries the argument

Meta-trained dual-branch neural network that outputs basis functions from Gaussian random fields, which are then used to build the collocation matrix for the linear solve.

Load-bearing premise

Basis functions learned from Gaussian random fields will generalize without retraining to the specific operators and boundary conditions of new PDEs.

What would settle it

A new PDE whose solution cannot be represented accurately by the fixed basis functions, producing large residual errors in the collocation least-squares solve even after the meta-training step.

Figures

Figures reproduced from arXiv: 2605.12368 by Zichuan Yang.

Figure 1
Figure 1. Figure 1: shows the full pipeline. At test time, we freeze the network. We compute exact derivatives with forward-mode automatic differentiation. We build a collocation matrix and solve it in one step. For non-linear PDEs, we apply the Newton–Raphson method. This gives fast and stable updates. We reach high accuracy in a few iterations. We test MetaColloc on six PDEs. It matches state-of-the-art accuracy on smooth p… view at source ↗
Figure 2
Figure 2. Figure 2: Frequency sweep exposes the spectral limitations of MetaColloc. Although the meta￾learned basis can represent high-frequency modes in function space, its Laplacian responses become increasingly unstable as frequency grows. The divergence between RMSE(u) and RMSE(∆u) highlights a structural gap between interpolation accuracy and PDE operator compatibility. This phenomenon reflects a general property of func… view at source ↗
Figure 3
Figure 3. Figure 3: RMSE as a function of the number of iterations [PITH_FULL_IMAGE:figures/full_fig_p017_3.png] view at source ↗
Figure 4
Figure 4. Figure 4: Exact solution, MetaColloc prediction, and absolute error heatmap for the Poisson equation [PITH_FULL_IMAGE:figures/full_fig_p018_4.png] view at source ↗
Figure 5
Figure 5. Figure 5: Exact solution, MetaColloc prediction, and absolute error heatmap for the Poisson equation [PITH_FULL_IMAGE:figures/full_fig_p019_5.png] view at source ↗
read the original abstract

Solving partial differential equations (PDEs) with machine learning typically requires training a new neural network for every new equation. This optimization is slow. We introduce MetaColloc. It is an optimization-free and data-free framework that removes this bottleneck completely. We decouple basis discovery from the solving process. We meta-train a dual-branch neural network on diverse Gaussian Random Fields. This offline process creates a universal dictionary of neural basis functions. At test time, we freeze the network. We solve the PDE by assembling a collocation matrix. We find the solution through a single linear least squares step. For non-linear PDEs, we apply the Newton-Raphson method to achieve fast quadratic convergence. Our experiments across six 2D and 3D PDEs show massive improvements. MetaColloc reaches state-of-the-art accuracy on smooth and non-linear problems. It also reduces test-time computation by several orders of magnitude. Finally, we provide a detailed frequency sweep analysis. This analysis reveals a critical mismatch between function approximation and operator stability at extremely high frequencies. This profound finding opens a clear path toward future operator-aware meta-learning.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

3 major / 2 minor

Summary. The paper introduces MetaColloc, an optimization-free, data-free framework for solving PDEs. A dual-branch neural network is meta-trained offline on diverse Gaussian random fields to produce a fixed dictionary of neural basis functions. At test time the network is frozen and each new PDE is solved by assembling a collocation matrix and performing a single linear least-squares step (Newton-Raphson for nonlinear cases). Experiments on six 2D/3D PDEs claim state-of-the-art accuracy together with orders-of-magnitude reduction in test-time cost; a frequency-sweep study is presented that identifies a mismatch between function approximation quality and operator stability at high frequencies.

Significance. If the reported accuracy and generalization hold, the work would provide a practical route to decoupling expensive basis discovery from per-instance PDE solving, yielding a reusable dictionary that enables near-instant linear solves. The frequency-sweep analysis supplies a concrete diagnostic for when such dictionaries remain reliable, which is a useful contribution to neural-operator and physics-informed methods.

major comments (3)
  1. [frequency sweep analysis] The central generalization claim—that bases meta-trained on isotropic GRFs suffice for arbitrary test PDE operators and boundary conditions without retraining—receives no supporting analysis. The frequency-sweep section already documents operator-stability mismatches; the same mismatch can appear at moderate frequencies when the target operator differs from GRF statistics, rendering the single-step least-squares solve inaccurate or ill-conditioned.
  2. [experiments] No error analysis, condition-number bounds, or approximation-error guarantees are given for the span of the learned basis functions. Without these, the assertion that the collocation matrix yields SOTA accuracy on smooth and nonlinear problems rests solely on unverified experimental assertions.
  3. [experiments] The manuscript provides no ablation on the number of basis functions, meta-training distribution statistics, or sensitivity to boundary-condition mismatch. These omissions make it impossible to assess whether the reported orders-of-magnitude speed-up is achieved while preserving the claimed accuracy.
minor comments (2)
  1. [method] The dual-branch architecture and the precise construction of the collocation matrix should be stated with explicit equations and pseudocode.
  2. [experiments] Tables reporting L2 or H1 errors, wall-clock times, and direct comparisons to PINN, DeepONet, and FNO baselines are needed to substantiate the SOTA claim.

Simulated Author's Rebuttal

3 responses · 0 unresolved

We thank the referee for the constructive and detailed feedback. The comments raise important points on generalization, theoretical support, and experimental completeness. We address each major comment below, indicating where revisions will be made to strengthen the manuscript.

read point-by-point responses
  1. Referee: [frequency sweep analysis] The central generalization claim—that bases meta-trained on isotropic GRFs suffice for arbitrary test PDE operators and boundary conditions without retraining—receives no supporting analysis. The frequency-sweep section already documents operator-stability mismatches; the same mismatch can appear at moderate frequencies when the target operator differs from GRF statistics, rendering the single-step least-squares solve inaccurate or ill-conditioned.

    Authors: We appreciate this observation. The generalization claim is supported empirically by results on six PDEs spanning linear/nonlinear operators and varied boundary conditions, all solved accurately with the fixed GRF-trained dictionary. The frequency-sweep analysis shows breakdown only at extremely high frequencies; our test cases remain in the moderate-frequency regime where stability holds. In revision we will add a spectral comparison between training GRF statistics and test operators, plus new experiments on PDEs whose operators deviate further from GRF spectra. A full theoretical characterization of the stability region remains future work. revision: partial

  2. Referee: [experiments] No error analysis, condition-number bounds, or approximation-error guarantees are given for the span of the learned basis functions. Without these, the assertion that the collocation matrix yields SOTA accuracy on smooth and nonlinear problems rests solely on unverified experimental assertions.

    Authors: We agree that formal approximation bounds and condition-number guarantees would strengthen the theoretical foundation. Deriving such results for meta-learned neural bases requires substantial additional analysis that lies beyond the scope of the present work, which focuses on the practical framework and empirical performance. In the revision we will report empirical condition numbers observed across all test cases and clarify that SOTA accuracy claims are empirical, backed by the reported numerical results. revision: partial

  3. Referee: [experiments] The manuscript provides no ablation on the number of basis functions, meta-training distribution statistics, or sensitivity to boundary-condition mismatch. These omissions make it impossible to assess whether the reported orders-of-magnitude speed-up is achieved while preserving the claimed accuracy.

    Authors: We acknowledge the importance of these ablations for assessing robustness. The revised manuscript will include a new ablation subsection reporting: (i) accuracy and solve time versus number of basis functions (32–256), (ii) sensitivity to GRF hyperparameters (length scale and variance), and (iii) performance under boundary-condition mismatch (e.g., Dirichlet-trained dictionary tested on Neumann or mixed conditions). These results will quantify the accuracy–speed trade-off. revision: yes

Circularity Check

0 steps flagged

No significant circularity: test-time solve is independent of meta-training loss

full rationale

The paper decouples meta-training (on GRFs to produce frozen basis functions) from test-time solution (standard linear least-squares collocation on the fixed dictionary, or Newton-Raphson for nonlinear cases). This separation means the reported accuracy and speed claims do not reduce by the paper's own equations to a quantity defined in terms of the meta-training objective. No self-citations, uniqueness theorems, ansatzes smuggled via prior work, or fitted parameters renamed as predictions appear in the provided text. The frequency-sweep analysis is presented as an empirical observation rather than a load-bearing derivation. This is the normal case of a self-contained method whose central claim rests on external validation rather than internal redefinition.

Axiom & Free-Parameter Ledger

1 free parameters · 1 axioms · 0 invented entities

The central claim rests on the assumption that Gaussian random fields supply a sufficiently rich training distribution for basis functions that transfer to arbitrary PDEs, plus standard neural-network training hyperparameters that are not enumerated.

free parameters (1)
  • meta-training hyperparameters
    Network architecture, learning rate schedule, and number of random fields are chosen during the offline meta-training stage.
axioms (1)
  • domain assumption Gaussian random fields provide sufficient diversity to learn basis functions that generalize to unseen PDE operators and domains
    The training distribution is stated as diverse GRFs; transfer to test PDEs is assumed without further justification in the abstract.

pith-pipeline@v0.9.0 · 5490 in / 1227 out tokens · 29662 ms · 2026-05-13T05:53:54.666646+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Lean theorems connected to this paper

Citations machine-checked in the Pith Canon. Every link opens the source theorem in the public Lean library.

Reference graph

Works this paper leans on

50 extracted references · 50 canonical work pages · 1 internal anchor

  1. [1]

    Universal physics transformers: A framework for efficiently scal- ing neural operators

    Benedikt Alkin, Andreas Fürst, Simon Schmid, Lukas Gruber, Markus Holzleitner, and Johannes Brandstetter. Universal physics transformers: A framework for efficiently scal- ing neural operators. In A. Globerson, L. Mackey, D. Belgrave, A. Fan, U. Paquet, J. Tomczak, and C. Zhang, editors,Advances in Neural Information Processing Sys- tems, volume 37, pages...

  2. [2]

    Theory of reproducing kernels.Transactions of the American mathemati- cal society, 68(3):337–404, 1950

    Nachman Aronszajn. Theory of reproducing kernels.Transactions of the American mathemati- cal society, 68(3):337–404, 1950

  3. [3]

    Physics-informed diffusion models

    Jan-Hendrik Bastek, WaiChing Sun, and Dennis Kochmann. Physics-informed diffusion models. InThe Thirteenth International Conference on Learning Representations, 2025. URL https://openreview.net/forum?id=tpYeermigp

  4. [4]

    PRDP: Progressively refined differentiable physics

    Kanishk Bhatia, Felix Koehler, and Nils Thuerey. PRDP: Progressively refined differentiable physics. InThe Thirteenth International Conference on Learning Representations, 2025. URL https://openreview.net/forum?id=9Fh0z1JmPU

  5. [5]

    Solving partial differential equations with point source based on physics- informed neural networks.Journal of Computational Physics, 2021

    Yuntian Chen et al. Solving partial differential equations with point source based on physics- informed neural networks.Journal of Computational Physics, 2021

  6. [6]

    Aakanksha Chowdhery, Sharan Narang, Jacob Devlin, Maarten Bosma, Gaurav Mishra, Adam Roberts, Paul Barham, Hyung Won Chung, Charles Sutton, Sebastian Gehrmann, Parker Schuh, Kensen Shi, Sasha Tsvyashchenko, Joshua Maynez, Abhishek Rao, Parker Barnes, Yi Tay, Noam Shazeer, Vinodkumar Prabhakaran, Emily Reif, Nan Du, Ben Hutchinson, Reiner Pope, James Bradb...

  7. [7]

    Kronecker-factored approximate cur- vature for physics-informed neural networks

    Felix Dangel, Johannes Müller, and Marius Zeinhofer. Kronecker-factored approximate cur- vature for physics-informed neural networks. In A. Globerson, L. Mackey, D. Belgrave, A. Fan, U. Paquet, J. Tomczak, and C. Zhang, editors,Advances in Neural Information Processing Systems, volume 37, pages 34582–34636. Curran Associates, Inc., 2024. doi: 10.52202/079...

  8. [8]

    Suchuan Dong and Zongyi Li. Local extreme learning machines and domain decomposition for solving linear and nonlinear partial differential equations.Computer Methods in Applied Mechanics and Engineering, 387:114129, 2021. doi: 10.1016/j.cma.2021.114129

  9. [9]

    Solving high frequency and multi-scale PDEs with gaussian processes

    Shikai Fang, Madison Cooley, Da Long, Shibo Li, Mike Kirby, and Shandian Zhe. Solving high frequency and multi-scale PDEs with gaussian processes. InThe Twelfth International Conference on Learning Representations, 2024. URL https://openreview.net/forum? id=q4AEBLHuA6

  10. [10]

    SINGER: Stochastic network graph evolving operator for high dimensional PDEs

    Mingquan Feng, Yixin Huang, Weixin Liao, Yuhong Liu, Yizhou Liu, and Junchi Yan. SINGER: Stochastic network graph evolving operator for high dimensional PDEs. InThe Thirteenth International Conference on Learning Representations, 2025. URL https://openreview. net/forum?id=wVADj7yKee

  11. [11]

    Randnet-parareal: a time-parallel pde solver using random neural networks

    Guglielmo Gattiglio, Lyudmila Grigoryeva, and Massimiliano Tamborrino. Randnet-parareal: a time-parallel pde solver using random neural networks. In A. Globerson, L. Mackey, D. Bel- grave, A. Fan, U. Paquet, J. Tomczak, and C. Zhang, editors,Advances in Neural Information Processing Systems, volume 37, pages 94993–95025. Curran Associates, Inc., 2024. doi...

  12. [12]

    The power of linear combinations: Learning with random convolutions, 2023

    Paul Gavrikov and Janis Keuper. The power of linear combinations: Learning with random convolutions, 2023. URLhttps://arxiv.org/abs/2301.11360

  13. [13]

    Brown University, 2006

    Minh Ha Quang.Reproducing kernel Hilbert spaces in learning theory. Brown University, 2006

  14. [14]

    Newton informed neural operator for solv- ing nonlinear partial differential equations

    Wenrui Hao, Xinliang Liu, and Yahong Yang. Newton informed neural operator for solv- ing nonlinear partial differential equations. In A. Globerson, L. Mackey, D. Belgrave, A. Fan, U. Paquet, J. Tomczak, and C. Zhang, editors,Advances in Neural Information Pro- cessing Systems, volume 37, pages 120832–120860. Curran Associates, Inc., 2024. doi: 10.52202/07...

  15. [15]

    Sobolev regularity of gaussian random fields.Journal of Functional Analysis, 286(3):110241, 2024

    Iain Henderson. Sobolev regularity of gaussian random fields.Journal of Functional Analysis, 286(3):110241, 2024

  16. [16]

    Poseidon: Efficient founda- tion models for pdes

    Maximilian Herde, Bogdan Raoni ´c, Tobias Rohner, Roger Käppeli, Roberto Moli- naro, Emmanuel de Bézenac, and Siddhartha Mishra. Poseidon: Efficient founda- tion models for pdes. In A. Globerson, L. Mackey, D. Belgrave, A. Fan, U. Paquet, J. Tomczak, and C. Zhang, editors,Advances in Neural Information Processing Sys- tems, volume 37, pages 72525–72624. C...

  17. [17]

    Extreme learning machine: theory and applications

    Guang-Bin Huang, Qin-Yu Zhu, and Chee-Kheong Siew. Extreme learning machine: theory and applications. InNeurocomputing, 2006

  18. [18]

    Diffusionpde: Gen- erative pde-solving under partial observation

    Jiahe Huang, Guandao Yang, Zichen Wang, and Jeong Joon Park. Diffusionpde: Gen- erative pde-solving under partial observation. In A. Globerson, L. Mackey, D. Belgrave, A. Fan, U. Paquet, J. Tomczak, and C. Zhang, editors,Advances in Neural Information Pro- cessing Systems, volume 37, pages 130291–130323. Curran Associates, Inc., 2024. doi: 10.52202/079017...

  19. [19]

    Dual cone gradient descent for training physics- informed neural networks

    Youngsik Hwang and Dong-Young Lim. Dual cone gradient descent for training physics- informed neural networks. In A. Globerson, L. Mackey, D. Belgrave, A. Fan, U. Pa- quet, J. Tomczak, and C. Zhang, editors,Advances in Neural Information Processing Sys- tems, volume 37, pages 98563–98595. Curran Associates, Inc., 2024. doi: 10.52202/ 079017-3128. URL https...

  20. [20]

    PIG: Physics-informed gaussians as adaptive parametric mesh representations

    Namgyu Kang, Jaemin Oh, Youngjoon Hong, and Eunbyung Park. PIG: Physics-informed gaussians as adaptive parametric mesh representations. InThe Thirteenth International Con- ference on Learning Representations, 2025. URL https://openreview.net/forum?id= y5B0ca4mjt

  21. [21]

    Fourier neural operator for parametric partial differential equations

    Zongyi Li, Nikola Borislavov Kovachki, Kamyar Azizzadenesheli, Burigede liu, Kaushik Bhattacharya, Andrew Stuart, and Anima Anandkumar. Fourier neural operator for parametric partial differential equations. InInternational Conference on Learning Representations, 2021. URLhttps://openreview.net/forum?id=c8P9NQVtmnO

  22. [22]

    Fourier neural operator with learned deformations for pdes on general geometries.Journal of Machine Learning Research, 24(388):1–26, 2023

    Zongyi Li, Daniel Zhengyu Huang, Burigede Liu, and Anima Anandkumar. Fourier neural operator with learned deformations for pdes on general geometries.Journal of Machine Learning Research, 24(388):1–26, 2023. URLhttp://jmlr.org/papers/v24/23-0064.html

  23. [23]

    Solving partial differential equations with random feature models.Communi- cations in Nonlinear Science and Numerical Simulation, page 109343, 2025

    Chunyang Liao. Solving partial differential equations with random feature models.Communi- cations in Nonlinear Science and Numerical Simulation, page 109343, 2025

  24. [24]

    ConFIG: Towards conflict-free training of physics informed neural networks

    Qiang Liu, Mengyu Chu, and Nils Thuerey. ConFIG: Towards conflict-free training of physics informed neural networks. InThe Thirteenth International Conference on Learning Representa- tions, 2025. URLhttps://openreview.net/forum?id=APojAzJQiq

  25. [25]

    Hou, and Max Tegmark

    Ziming Liu, Yixuan Wang, Sachin Vaidya, Fabian Ruehle, James Halverson, Marin Soljacic, Thomas Y . Hou, and Max Tegmark. KAN: Kolmogorov–arnold networks. InThe Thirteenth International Conference on Learning Representations, 2025. URL https://openreview. net/forum?id=Ozo7qJ5vZi

  26. [26]

    Decoupled weight decay regularization

    Ilya Loshchilov and Frank Hutter. Decoupled weight decay regularization. InInternational Conference on Learning Representations, 2019. URL https://openreview.net/forum? id=Bkg6RiCqY7

  27. [27]

    Learning nonlinear operators via DeepONet based on the universal approximation theorem of operators.Nature Machine Intelligence, 3(3), 218–229 (2021)

    Lu Lu, Pengzhan Jin, Guofei Pang, Zhongqiang Zhang, and George Em Karniadakis. Learning nonlinear operators via deeponet based on the universal approximation theorem of operators. Nature machine intelligence, 3(3):218–229, 2021. doi: 10.1038/s42256-021-00302-5

  28. [28]

    A multiple transferable neural network method with domain decomposition for elliptic interface problems.Journal of Computational Physics, 530:113902, 2025

    Tianzheng Lu, Lili Ju, and Liyong Zhu. A multiple transferable neural network method with domain decomposition for elliptic interface problems.Journal of Computational Physics, 530:113902, 2025. ISSN 0021-9991. doi: https://doi.org/10.1016/j.jcp.2025.113902. URL https://www.sciencedirect.com/science/article/pii/S0021999125001858

  29. [29]

    Stochastic processes with sample paths in reproducing kernel hilbert spaces.Transactions of the American Mathematical Society, 353(10):3945–3969, 2001

    Milan Luki´c and Jay Beder. Stochastic processes with sample paths in reproducing kernel hilbert spaces.Transactions of the American Mathematical Society, 353(10):3945–3969, 2001

  30. [30]

    Multiple physics pre- training for spatiotemporal surrogate models

    Michael McCabe, Bruno Régaldo-Saint Blancard, Liam Parker, Ruben Ohana, Miles Cran- mer, Alberto Bietti, Michael Eickenberg, Siavash Golkar, Geraud Krawezik, Francois Lanusse, Mariel Pettee, Tiberiu Tesileanu, Kyunghyun Cho, and Shirley Ho. Multiple physics pre- training for spatiotemporal surrogate models. In A. Globerson, L. Mackey, D. Belgrave, A. Fan,...

  31. [31]

    On the spectral bias of neural networks

    Nasim Rahaman, Aristide Baratin, Devansh Arpit, Felix Eberhardt, Bengio Yoshua, and Aaron Courville. On the spectral bias of neural networks. InInternational Conference on Machine Learning, pages 5301–5310. PMLR, 2019. 12

  32. [32]

    Random features for large-scale kernel machines

    Ali Rahimi and Benjamin Recht. Random features for large-scale kernel machines. InAdvances in neural information processing systems, 2007

  33. [33]

    Yeh, Jean Kossaifi, Kamyar Az- izzadenesheli, and Anima Anandkumar

    Ashiqur Rahman, Robert Joseph George, Mogab Elleithy, Daniel Leibovici, Zongyi Li, Boris Bonev, Colin White, Julius Berner, Raymond A. Yeh, Jean Kossaifi, Kamyar Az- izzadenesheli, and Anima Anandkumar. Pretraining codomain attention neural operators for solving multiphysics pdes. In A. Globerson, L. Mackey, D. Belgrave, A. Fan, U. Pa- quet, J. Tomczak, a...

  34. [34]

    Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations

    M. Raissi, P. Perdikaris, and G.E. Karniadakis. Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations.Journal of Computational Physics, 378:686–707, 2019. ISSN 0021-9991. doi: https://doi.org/10.1016/j.jcp.2018.10.045. URL https://www.sciencedirect.com/ scien...

  35. [35]

    Aroma: Preserving spatial structure for latent pde modeling with lo- cal neural fields

    Louis Serrano, Thomas X Wang, Etienne Le Naour, Jean-Noël Vittaut, and Patrick Gallinari. Aroma: Preserving spatial structure for latent pde modeling with lo- cal neural fields. In A. Globerson, L. Mackey, D. Belgrave, A. Fan, U. Paquet, J. Tomczak, and C. Zhang, editors,Advances in Neural Information Processing Sys- tems, volume 37, pages 13489–13521. Cu...

  36. [36]

    Improved sampling of diffusion models in fluid dynamics with tweedie’s formula

    Youssef Shehata, Benjamin Holzschuh, and Nils Thuerey. Improved sampling of diffusion models in fluid dynamics with tweedie’s formula. InThe Thirteenth International Conference on Learning Representations, 2025. URL https://openreview.net/forum?id=0FbzC7B9xI

  37. [37]

    Turner, and Emile Mathieu

    Aliaksandra Shysheya, Cristiana Diaconu, Federico Bergamin, Paris Perdikaris, José Miguel Hernández-Lobato, Richard E. Turner, and Emile Mathieu. On conditional diffusion mod- els for pde simulations. In A. Globerson, L. Mackey, D. Belgrave, A. Fan, U. Pa- quet, J. Tomczak, and C. Zhang, editors,Advances in Neural Information Processing Sys- tems, volume ...

  38. [38]

    Fourier fea- tures let networks learn high frequency functions in low dimensional domains

    Matthew Tancik, Pratul Srinivasan, Ben Mildenhall, Sara Fridovich-Keil, Nithin Ragha- van, Utkarsh Singhal, Ravi Ramamoorthi, Jonathan Barron, and Ren Ng. Fourier fea- tures let networks learn high frequency functions in low dimensional domains. In H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin, editors,Advances in Neural Information Proce...

  39. [39]

    Metamizer: A versatile neural optimizer for fast and accurate physics simulations

    Nils Wandel, Stefan Schulz, and Reinhard Klein. Metamizer: A versatile neural optimizer for fast and accurate physics simulations. InThe Thirteenth International Conference on Learning Representations, 2025. URLhttps://openreview.net/forum?id=60TXv9Xif5

  40. [40]

    Understanding and mitigating gradient flow pathologies in physics-informed neural networks

    Sifan Wang, Yujun Teng, and Paris Perdikaris. Understanding and mitigating gradient flow pathologies in physics-informed neural networks. InSIAM Journal on Scientific Computing, 2021

  41. [41]

    On the eigenvector bias of fourier feature networks: From regression to solving multi-scale pdes with physics-informed neural networks

    Sifan Wang, Hanwen Wang, and Paris Perdikaris. On the eigenvector bias of fourier feature networks: From regression to solving multi-scale pdes with physics-informed neural networks. InComputer Methods in Applied Mechanics and Engineering, 2021

  42. [42]

    Respecting causality is all you need for training physics-informed neural networks

    Sifan Wang, Shyam Sankaran, and Paris Perdikaris. Respecting causality is all you need for training physics-informed neural networks. InarXiv preprint arXiv:2203.07404, 2022. 13

  43. [43]

    Pappas, and Paris Perdikaris

    Sifan Wang, Jacob H Seidman, Shyam Sankaran, Hanwen Wang, George J. Pappas, and Paris Perdikaris. CVit: Continuous vision transformer for operator learning. InThe Thirteenth International Conference on Learning Representations, 2025. URL https://openreview. net/forum?id=cRnCcuLvyr

  44. [44]

    Latent neural operator for solving forward and in- verse pde problems

    Tian Wang and Chuang Wang. Latent neural operator for solving forward and in- verse pde problems. In A. Globerson, L. Mackey, D. Belgrave, A. Fan, U. Paquet, J. Tomczak, and C. Zhang, editors,Advances in Neural Information Processing Sys- tems, volume 37, pages 33085–33107. Curran Associates, Inc., 2024. doi: 10.52202/ 079017-1042. URL https://proceedings...

  45. [45]

    CL-diffphycon: Closed-loop diffusion control of complex physical systems

    Long Wei, Haodong Feng, Yuchen Yang, Ruiqi Feng, Peiyan Hu, Xiang Zheng, Tao Zhang, Dixia Fan, and Tailin Wu. CL-diffphycon: Closed-loop diffusion control of complex physical systems. InThe Thirteenth International Conference on Learning Representations, 2025. URL https://openreview.net/forum?id=PiHGrTTnvb

  46. [46]

    MIT press Cambridge, MA, 2006

    Christopher KI Williams and Carl Edward Rasmussen.Gaussian processes for machine learning, volume 2. MIT press Cambridge, MA, 2006

  47. [47]

    Ropinn: Re- gion optimized physics-informed neural networks

    Haixu Wu, Huakun Luo, Yuezhou Ma, Jianmin Wang, and Mingsheng Long. Ropinn: Re- gion optimized physics-informed neural networks. In A. Globerson, L. Mackey, D. Belgrave, A. Fan, U. Paquet, J. Tomczak, and C. Zhang, editors,Advances in Neural Information Pro- cessing Systems, volume 37, pages 110494–110532. Curran Associates, Inc., 2024. doi: 10.52202/0790...

  48. [48]

    Transferable neural networks for partial differential equations.Journal of Scientific Computing, 99(2):31, 2024

    Zezhong Zhang, Feng Bao, Lili Ju, and Guannan Zhang. Transferable neural networks for partial differential equations.Journal of Scientific Computing, 99(2):31, 2024. doi: 10.1007/ s10915-024-02463-y

  49. [49]

    Alias- free mamba neural operator

    Jianwei Zheng, Wei Li, Ni Xu, Junwei Zhu, Xiaoxu Lin, and Xiaoqin Zhang. Alias- free mamba neural operator. In A. Globerson, L. Mackey, D. Belgrave, A. Fan, U. Pa- quet, J. Tomczak, and C. Zhang, editors,Advances in Neural Information Processing Sys- tems, volume 37, pages 52962–52995. Curran Associates, Inc., 2024. doi: 10.52202/ 079017-1678. URL https:/...

  50. [50]

    FP32+FP64

    Anthony Zhou, Zijie Li, Michael Schneier, John R Buchanan Jr, and Amir Barati Farimani. Text2PDE: Latent diffusion models for accessible physics simulation. InThe Thirteenth International Conference on Learning Representations, 2025. URL https://openreview. net/forum?id=Nb3a8aUGfj. 14 A Experimental Setup and Hyperparameters We provide the exact settings ...