pith. machine review for the scientific record. sign in

arxiv: 2202.11214 · v1 · submitted 2022-02-22 · ⚛️ physics.ao-ph · cs.LG

Recognition: 3 theorem links

· Lean Theorem

FourCastNet: A Global Data-driven High-resolution Weather Model using Adaptive Fourier Neural Operators

Animashree Anandkumar, Ashesh Chattopadhyay, David Hall, Jaideep Pathak, Kamyar Azizzadenesheli, Karthik Kashinath, Morteza Mardani, Pedram Hassanzadeh, Peter Harrington, Sanjeev Raja, Shashank Subramanian, Thorsten Kurth, Zongyi Li

Pith reviewed 2026-05-12 09:59 UTC · model grok-4.3

classification ⚛️ physics.ao-ph cs.LG
keywords weather forecastingdata-driven modelingneural operatorsglobal high-resolution predictionensemble forecastingprecipitationextreme weather
0
0 comments X

The pith

FourCastNet matches the ECMWF IFS forecasting accuracy at short lead times for large-scale variables while outperforming it on fine-scale features like precipitation and generating a week-long global forecast in under two seconds.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper introduces FourCastNet as a data-driven global weather model that delivers accurate short- to medium-range predictions at 0.25 degree resolution. It establishes that this approach matches the accuracy of the leading numerical weather prediction system, the ECMWF IFS, for broad atmospheric patterns at short lead times and exceeds it for detailed variables such as precipitation. The model runs a full week forecast in less than two seconds, enabling large numbers of ensemble members that improve estimates of extreme event probabilities. This speed and accuracy combination matters for applications like wind energy planning and tropical cyclone prediction because it allows rapid, low-cost exploration of many possible futures.

Core claim

FourCastNet is a global data-driven weather forecasting model that provides accurate short to medium-range global predictions at 0.25° resolution using adaptive Fourier neural operators. It accurately forecasts high-resolution, fast-timescale variables such as the surface wind speed, precipitation, and atmospheric water vapor. FourCastNet matches the forecasting accuracy of the ECMWF Integrated Forecasting System (IFS) at short lead times for large-scale variables, while outperforming IFS for variables with complex fine-scale structure, including precipitation. FourCastNet generates a week-long forecast in less than 2 seconds, orders of magnitude faster than IFS.

What carries the argument

Adaptive Fourier Neural Operators that learn resolution-invariant mappings between atmospheric state functions to enable efficient high-resolution forecasting.

If this is right

  • Enables creation of rapid and inexpensive large-ensemble forecasts with thousands of members to improve probabilistic forecasting.
  • Supports better prediction and planning for extreme weather events such as tropical cyclones, extra-tropical cyclones, and atmospheric rivers.
  • Provides accurate surface wind speed forecasts that aid planning of wind energy resources.
  • Acts as a valuable addition to the meteorology toolkit that can augment rather than replace traditional NWP models.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The approach could be hybridized with physics-based constraints to enforce conservation laws and improve stability at longer lead times.
  • Low computational cost opens possibilities for real-time ensemble nowcasting at scales not feasible with IFS.
  • The same operator framework might accelerate ensemble climate projections by running far more members than current physics models allow.

Load-bearing premise

The neural operator trained on historical reanalysis data will generalize accurately to future unseen weather states, including rare extreme events, without explicit enforcement of physical conservation laws or stability constraints.

What would settle it

A side-by-side evaluation on an independent extreme precipitation event or tropical cyclone where FourCastNet errors exceed those of IFS at the same lead time would show the generalization claim does not hold.

read the original abstract

FourCastNet, short for Fourier Forecasting Neural Network, is a global data-driven weather forecasting model that provides accurate short to medium-range global predictions at $0.25^{\circ}$ resolution. FourCastNet accurately forecasts high-resolution, fast-timescale variables such as the surface wind speed, precipitation, and atmospheric water vapor. It has important implications for planning wind energy resources, predicting extreme weather events such as tropical cyclones, extra-tropical cyclones, and atmospheric rivers. FourCastNet matches the forecasting accuracy of the ECMWF Integrated Forecasting System (IFS), a state-of-the-art Numerical Weather Prediction (NWP) model, at short lead times for large-scale variables, while outperforming IFS for variables with complex fine-scale structure, including precipitation. FourCastNet generates a week-long forecast in less than 2 seconds, orders of magnitude faster than IFS. The speed of FourCastNet enables the creation of rapid and inexpensive large-ensemble forecasts with thousands of ensemble-members for improving probabilistic forecasting. We discuss how data-driven deep learning models such as FourCastNet are a valuable addition to the meteorology toolkit to aid and augment NWP models.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 2 minor

Summary. The paper introduces FourCastNet, a global data-driven weather forecasting model based on Adaptive Fourier Neural Operators (AFNO) trained on ERA5 reanalysis. It claims accurate 0.25° resolution short-to-medium range forecasts that match ECMWF IFS accuracy for large-scale variables at short lead times, outperform IFS on fine-scale variables including precipitation, and enable week-long forecasts in under 2 seconds, with implications for ensemble forecasting and extreme weather prediction.

Significance. If the results hold under rigorous validation, this work establishes that neural operator models can achieve competitive skill with operational NWP systems without explicit physics constraints, while offering orders-of-magnitude speedups that enable large ensembles. This positions data-driven approaches as a practical complement to traditional models for probabilistic forecasting and rapid scenario generation.

major comments (2)
  1. [§4 (Results)] §4 (Results): Performance comparisons to IFS report RMSE and ACC metrics but omit details on the exact held-out validation years from ERA5, statistical error bars or significance tests on the metrics, and explicit checks for distribution shift. This directly affects the load-bearing claim that the model matches or exceeds IFS on held-out data for key variables.
  2. [§3 (Methods) and §4.3] §3 (Methods) and §4.3: No verification is provided that forecasts conserve quantities such as total column water vapor, mass, or energy to within observational uncertainty, nor are there tests on out-of-distribution extremes (e.g., record events after the training cutoff). These omissions are load-bearing for the generalization assumption underlying multi-day accuracy claims.
minor comments (2)
  1. [Abstract] Abstract: Mentions implications for atmospheric rivers and tropical cyclones but the results do not include dedicated metrics or case studies for these phenomena.
  2. [Figures] Figure captions and legends: Some figures comparing FourCastNet and IFS forecasts would benefit from explicit lead-time annotations and variable units to improve clarity.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for their constructive feedback, which has helped us identify areas to improve the clarity and rigor of our manuscript. Below, we provide point-by-point responses to the major comments.

read point-by-point responses
  1. Referee: [§4 (Results)] Performance comparisons to IFS report RMSE and ACC metrics but omit details on the exact held-out validation years from ERA5, statistical error bars or significance tests on the metrics, and explicit checks for distribution shift. This directly affects the load-bearing claim that the model matches or exceeds IFS on held-out data for key variables.

    Authors: We thank the referee for highlighting these omissions. The manuscript specifies in Section 3 that the model is trained on ERA5 data from 1979 to 2017 and evaluated on 2018-2020. We will revise the results section to explicitly state the validation years, include statistical error bars (e.g., via bootstrapping) on the reported RMSE and ACC values, perform significance tests (e.g., paired t-tests) to compare FourCastNet and IFS, and add checks for distribution shift by comparing means and variances of key variables between training and validation periods. These additions will be incorporated in the revised manuscript. revision: yes

  2. Referee: [§3 (Methods) and §4.3] No verification is provided that forecasts conserve quantities such as total column water vapor, mass, or energy to within observational uncertainty, nor are there tests on out-of-distribution extremes (e.g., record events after the training cutoff). These omissions are load-bearing for the generalization assumption underlying multi-day accuracy claims.

    Authors: We agree that verifying physical consistency and robustness to extremes is important. Although FourCastNet is data-driven and does not hard-code conservation, we will add in the revised paper an evaluation of approximate conservation by tracking the evolution of integrated quantities like total column water vapor, total mass, and energy over forecast lead times and comparing deviations to ERA5 observational uncertainties. For out-of-distribution extremes, we will include additional analysis of performance on record events in the test period (post-2017), such as the most extreme precipitation events or temperature records in 2018-2020. This will be added to Section 4.3 to support the generalization claims. revision: yes

Circularity Check

0 steps flagged

No circularity detected; claims rest on empirical training and external benchmark comparison.

full rationale

The paper describes a data-driven AFNO-based neural network trained on ERA5 reanalysis and evaluated via direct RMSE/ACC metrics against held-out years and the independent ECMWF IFS model. No derivation chain reduces predictions to fitted parameters by construction, no self-citation load-bearing uniqueness theorems are invoked to force the architecture, and performance claims are falsifiable against external operational forecasts rather than tautological. Generalization risk to extremes is a standard ML limitation but does not create circularity in the reported results.

Axiom & Free-Parameter Ledger

1 free parameters · 1 axioms · 0 invented entities

The central performance claims rest on the assumption that patterns learned from past reanalysis data will transfer to future states. No new physical entities are postulated; the model is purely data-driven.

free parameters (1)
  • neural network weights
    Learned parameters of the adaptive Fourier neural operator fitted to ERA5 reanalysis data.
axioms (1)
  • domain assumption Historical reanalysis data sufficiently represents the distribution of future atmospheric states for the forecast horizons considered
    Implicit in training on past data and evaluating generalization to later periods.

pith-pipeline@v0.9.0 · 5563 in / 1215 out tokens · 32110 ms · 2026-05-12T09:59:09.995006+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Lean theorems connected to this paper

Citations machine-checked in the Pith Canon. Every link opens the source theorem in the public Lean library.

Forward citations

Cited by 33 Pith papers

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. Convergent Stochastic Training of Attention and Understanding LoRA

    cs.LG 2026-05 unverdicted novelty 8.0

    Attention and LoRA regression losses induce Poincaré inequalities under mild regularization, so SGD-mimicking SDEs converge to minimizers with no assumptions on data or model size.

  2. Stable Attention Response for Reliable Precipitation Nowcasting

    cs.LG 2026-05 conditional novelty 7.0

    HARECast stabilizes cross-sample variance in attention-response energy via group-wise regularization to reduce prediction errors in precipitation nowcasting.

  3. WindINR: Latent-State INR for Fast Local Wind Query and Correction in Complex Terrain

    cs.AI 2026-05 unverdicted novelty 7.0

    WindINR achieves continuous high-resolution local wind queries and sparse-observation correction in complex terrain by updating only a compact latent state, delivering 2.6x speedup over full-network fine-tuning in OSS...

  4. CATO: Charted Attention for Neural PDE Operators

    cs.AI 2026-05 unverdicted novelty 7.0

    CATO learns a continuous latent chart for efficient axial attention on PDE meshes and adds derivative-aware supervision to improve accuracy and reduce oversmoothing on general geometries.

  5. Controlling Transient Amplification Improves Long-horizon Rollouts

    cs.LG 2026-05 unverdicted novelty 7.0

    Commutativity regularization on Jacobians reduces transient error amplification in neural simulators, enabling stable rollouts over thousands of steps on physical and climate data.

  6. QuadNorm: Resolution-Robust Normalization for Neural Operators

    cs.LG 2026-05 unverdicted novelty 7.0

    QuadNorm uses quadrature-based moments instead of uniform averaging in normalization layers, achieving O(h²) consistency across resolutions and better cross-resolution transfer in neural operators.

  7. GPROF-IR: An Improved Single-Channel Infrared Precipitation Retrieval for Merged Satellite Precipitation Products

    physics.ao-ph 2026-05 unverdicted novelty 7.0

    GPROF-IR is a CNN-based retrieval that uses temporal context in geostationary IR observations to produce precipitation estimates with lower error than prior IR methods and climatological consistency with PMW retrieval...

  8. PODiff: Latent Diffusion in Proper Orthogonal Decomposition Space for Scientific Super-Resolution

    cs.LG 2026-05 unverdicted novelty 7.0

    PODiff performs conditional diffusion in a fixed, variance-ordered POD latent space to enable efficient probabilistic super-resolution of high-dimensional scientific fields with lower memory and better-calibrated unce...

  9. Faster by Design: Interactive Aerodynamics via Neural Surrogates Trained on Expert-Validated CFD

    cs.LG 2026-04 unverdicted novelty 7.0

    A graph-based neural operator trained on expert-validated race-car CFD data reaches accuracy levels usable for early-stage interactive aerodynamic design exploration.

  10. Conflated Inverse Modeling to Generate Diverse and Temperature-Change Inducing Urban Vegetation Patterns

    cs.CV 2026-04 unverdicted novelty 7.0

    A diffusion generative inverse model conditioned on temperature targets produces diverse, physically plausible urban vegetation patterns that achieve specified regional temperature shifts.

  11. McCast: Memory-Guided Latent Drift Correction for Long-Horizon Precipitation Nowcasting

    cs.LG 2026-05 unverdicted novelty 6.0

    McCast uses a Drift-Corrective Memory Bank to actively correct latent drift in autoregressive precipitation nowcasting for more coherent long-horizon forecasts.

  12. Multi-Quantile Regression for Extreme Precipitation Downscaling

    cs.LG 2026-05 unverdicted novelty 6.0

    Q-SRDRN multi-quantile network with pinball loss and per-quantile heads detects extreme precipitation events up to 18 times more effectively than deterministic baselines while preserving augmentation benefits for the median.

  13. UFO: A Domain-Unification-Free Operator Framework for Generalized Operator Learning

    cs.LG 2026-05 unverdicted novelty 6.0

    UFO is a cross-domain neural operator framework that achieves discretization decoupling via adaptive jointly-conditioned interactions among distinct domain representations.

  14. Don't Fix the Basis -- Learn It: Spectral Representation with Adaptive Basis Learning for PDEs

    cs.LG 2026-05 unverdicted novelty 6.0

    ABLE learns a spatially adaptive Parseval frame from data via an ancillary density to replace fixed bases in spectral neural operators for PDEs.

  15. PnP-Corrector: A Universal Correction Framework for Coupled Spatiotemporal Forecasting

    cs.AI 2026-05 unverdicted novelty 6.0

    PnP-Corrector decouples physics simulation from error correction to counter reciprocal error amplification in coupled spatiotemporal forecasting, cutting error by 29% in a 300-day ocean-atmosphere test.

  16. PnP-Corrector: A Universal Correction Framework for Coupled Spatiotemporal Forecasting

    cs.AI 2026-05 unverdicted novelty 6.0

    PnP-Corrector decouples physics simulation from error correction via a plug-and-play agent, cutting error by 29% in 300-day global ocean-atmosphere forecasts.

  17. WeatherSyn: An Instruction Tuning MLLM For Weather Forecasting Report Generation

    cs.CL 2026-05 unverdicted novelty 6.0

    WeatherSyn is the first instruction-tuned MLLM for weather forecasting report generation, outperforming closed-source models on a new dataset of 31 US cities across 8 weather aspects.

  18. Sparse Random-Feature Neural Networks with Krylov-Based SVD for Singularly Perturbed ODE

    math.NA 2026-05 unverdicted novelty 6.0

    Sparse RFNNs with sSVD via Lanczos-Golub-Kahan bidiagonalization maintain accuracy while improving efficiency and robustness for 1D steady convection-diffusion equations with strong advection.

  19. Tyche: One Step Flow for Efficient Probabilistic Weather Forecasting

    cs.LG 2026-05 unverdicted novelty 6.0

    Tyche achieves competitive probabilistic weather forecasting skill and calibration using a single-step flow model with JVP-regularized training and rollout finetuning.

  20. Towards accurate extreme event likelihoods from diffusion model climate emulators

    physics.ao-ph 2026-05 unverdicted novelty 6.0

    Diffusion model climate emulators provide probability density estimates that allow likelihood calculations and odds-ratio-based importance sampling for extreme events such as tropical cyclones.

  21. M-CaStLe: Uncovering Local Causal Structures in Multivariate Space-Time Gridded Data

    cs.LG 2026-05 unverdicted novelty 6.0

    M-CaStLe generalizes local stencil-based causal discovery to the multivariate case and decomposes resulting graphs into reaction and spatial components for interpretation in space-time gridded data.

  22. Skillful Global Ocean Emulation and the Role of Correlation-Aware Loss

    physics.ao-ph 2026-04 unverdicted novelty 6.0

    A GraphCast-based ocean emulator achieves skillful 10-15 day forecasts, with a Mahalanobis loss that accounts for variable correlations improving performance over MSE and acting as a statistical-dynamical regularizer.

  23. Earth System Foundation Model (ESFM): A unified framework for heterogeneous data integration and forecasting

    physics.ao-ph 2026-04 unverdicted novelty 6.0

    ESFM is a single open foundation model that unifies heterogeneous Earth data sources and forecasts missing regions while preserving inter-variable physical relationships.

  24. Global Attention with Linear Complexity for Exascale Generative Data Assimilation in Earth System Prediction

    cs.LG 2026-04 unverdicted novelty 6.0

    STORM achieves linear-complexity global attention in a generative DA framework, scaling to 20 billion tokens and 1.6 ExaFLOPs on 32k GPUs for km-scale Earth modeling.

  25. Neural Stochastic Processes for Satellite Precipitation Refinement

    cs.CV 2026-04 unverdicted novelty 6.0

    NSP model fuses satellite and gauge data with neural processes and SDEs, outperforming 13 baselines and JAXA's operational product on a new 43k-sample US benchmark across six metrics.

  26. A PMP-inspired Evaluation Framework for Assessing Deep-Learning Earth System Models

    physics.ao-ph 2026-04 unverdicted novelty 6.0

    The paper presents a PMP-based evaluation framework to test deep-learning Earth system models on climatology and modes of variability using observational data.

  27. Neural Operators for Multi-Task Control and Adaptation

    cs.LG 2026-04 unverdicted novelty 6.0

    Neural operators approximate the solution operator for multi-task optimal control, generalizing to new tasks and enabling efficient adaptation via branch-trunk structure and meta-training.

  28. A Multimodal Vision Transformer-based Modeling Framework for Prediction of Fluid Flows in Energy Systems

    physics.flu-dyn 2026-04 unverdicted novelty 6.0

    A multimodal SwinV2-UNet vision transformer conditioned on data modality and time predicts spatiotemporal fluid flows and reconstructs unobserved fields from limited views using CFD data of argon jet injection.

  29. QuantWeather: Quantile-Aware Probabilistic Forecasting for Subseasonal Precipitation

    cs.CE 2026-05 unverdicted novelty 5.0

    QuantWeather is an end-to-end dual-head neural network that produces calibrated quantile-based probabilistic forecasts for subseasonal precipitation, achieving higher skill and lower inference cost than ensemble metho...

  30. PINN-Cast: Exploring the Role of Continuous-Depth NODE in Transformers and Physics Informed Loss as Soft Physical Constraints in Short-term Weather Forecasting

    cs.LG 2026-04 unverdicted novelty 5.0

    PINN-Cast combines continuous-depth Neural ODEs inside transformer blocks with a two-branch attention module and physics-informed loss to produce short-term weather forecasts that respect governing physical principles.

  31. PDE-regularized Dynamics-informed Diffusion with Uncertainty-aware Filtering for Long-Horizon Dynamics

    cs.LG 2026-04 unverdicted novelty 5.0

    PDYffusion combines PDE-regularized diffusion interpolation with UKF-based uncertainty-aware forecasting to deliver more stable and accurate long-horizon dynamical predictions than standard approaches.

  32. Regimes of Scale in AI Meteorology

    cs.HC 2026-04 unverdicted novelty 5.0

    AI/ML weather tools face integration challenges from mismatched 'regimes of scale' in how data and models are organized compared to traditional meteorology practices.

  33. What if AI systems weren't chatbots?

    cs.CY 2026-05 unverdicted novelty 3.0

    Chatbot AI systems often fail complex needs while projecting authority, contributing to deskilling, labor displacement, economic concentration, and high environmental costs, so alternative pluralistic and task-specifi...

Reference graph

Works this paper leans on

38 extracted references · 38 canonical work pages · cited by 32 Pith papers

  1. [1]

    The quiet revolution of numerical weather prediction

    Peter Bauer, Alan Thorpe, and Gilbert Brunet. The quiet revolution of numerical weather prediction. Nature, 525 0 (7567): 0 47--55, 2015

  2. [2]

    Advances in weather prediction

    Richard B Alley, Kerry A Emanuel, and Fuqing Zhang. Advances in weather prediction. Science, 363 0 (6425): 0 342--344, 2019

  3. [3]

    Weather prediction by numerical process

    Lewis Fry Richardson. Weather prediction by numerical process. Cambridge university press, 2007

  4. [4]

    Can deep learning beat numerical weather prediction? Philosophical Transactions of the Royal Society A, 379 0 (2194): 0 20200097, 2021

    MG Schultz, C Betancourt, B Gong, F Kleinert, M Langguth, LH Leufen, Amirpasha Mozaffari, and S Stadtler. Can deep learning beat numerical weather prediction? Philosophical Transactions of the Royal Society A, 379 0 (2194): 0 20200097, 2021

  5. [5]

    Climbing down charney’s ladder: machine learning and the post-dennard era of computational climate science

    V Balaji. Climbing down charney’s ladder: machine learning and the post-dennard era of computational climate science. Philosophical Transactions of the Royal Society A, 379 0 (2194): 0 20200085, 2021

  6. [6]

    Towards neural E arth system modelling by integrating artificial intelligence in E arth system science

    Christopher Irrgang, Niklas Boers, Maike Sonnewald, Elizabeth A Barnes, Christopher Kadow, Joanna Staneva, and Jan Saynisch-Wagner. Towards neural E arth system modelling by integrating artificial intelligence in E arth system science. Nature Machine Intelligence, 3 0 (8): 0 667--674, 2021

  7. [7]

    Deep learning and process understanding for data-driven earth system science

    Markus Reichstein, Gustau Camps-Valls, Bjorn Stevens, Martin Jung, Joachim Denzler, Nuno Carvalhais, et al. Deep learning and process understanding for data-driven earth system science. Nature, 566 0 (7743): 0 195--204, 2019

  8. [8]

    Predicting weather forecast uncertainty with machine learning

    Sebastian Scher and Gabriele Messori. Predicting weather forecast uncertainty with machine learning. Quarterly Journal of the Royal Meteorological Society, 144 0 (717): 0 2830--2841, 2018

  9. [9]

    Weather and climate forecasting with neural networks: using general circulation models (gcms) with different complexity as a study ground

    Sebastian Scher and Gabriele Messori. Weather and climate forecasting with neural networks: using general circulation models (gcms) with different complexity as a study ground. Geoscientific Model Development, 12 0 (7): 0 2797--2809, 2019

  10. [10]

    Analog forecasting of extreme-causing weather patterns using deep learning

    Ashesh Chattopadhyay, Ebrahim Nabizadeh, and Pedram Hassanzadeh. Analog forecasting of extreme-causing weather patterns using deep learning. Journal of Advances in Modeling Earth Systems, 12 0 (2): 0 e2019MS001958, 2020 a

  11. [11]

    Can machines learn to predict weather? using deep learning to predict gridded 500-hpa geopotential height from historical weather data

    Jonathan A Weyn, Dale R Durran, and Rich Caruana. Can machines learn to predict weather? using deep learning to predict gridded 500-hpa geopotential height from historical weather data. Journal of Advances in Modeling Earth Systems, 11 0 (8): 0 2680--2693, 2019

  12. [12]

    Improving data-driven global weather prediction using deep convolutional neural networks on a cubed sphere

    Jonathan A Weyn, Dale R Durran, and Rich Caruana. Improving data-driven global weather prediction using deep convolutional neural networks on a cubed sphere. Journal of Advances in Modeling Earth Systems, 12 0 (9): 0 e2020MS002109, 2020

  13. [13]

    Sub-seasonal forecasting with a large ensemble of deep-learning weather prediction models

    Jonathan A Weyn, Dale R Durran, Rich Caruana, and Nathaniel Cresswell-Clay. Sub-seasonal forecasting with a large ensemble of deep-learning weather prediction models. arXiv preprint arXiv:2102.05107, 2021

  14. [14]

    Weatherbench: a benchmark data set for data-driven weather forecasting

    Stephan Rasp, Peter D Dueben, Sebastian Scher, Jonathan A Weyn, Soukayna Mouatadid, and Nils Thuerey. Weatherbench: a benchmark data set for data-driven weather forecasting. Journal of Advances in Modeling Earth Systems, 12 0 (11): 0 e2020MS002203, 2020

  15. [15]

    Data-driven medium-range weather prediction with a resnet pretrained on climate simulations: A new model for weatherbench

    Stephan Rasp and Nils Thuerey. Data-driven medium-range weather prediction with a resnet pretrained on climate simulations: A new model for weatherbench. Journal of Advances in Modeling Earth Systems, 13 0 (2): 0 e2020MS002405, 2021 a

  16. [16]

    Purely data-driven medium-range weather forecasting achieves comparable skill to physical models at similar resolution

    Stephan Rasp and Nils Thuerey. Purely data-driven medium-range weather forecasting achieves comparable skill to physical models at similar resolution. arXiv preprint arXiv:2008.08626, 2020

  17. [17]

    Towards physically consistent data-driven weather forecasting: Integrating data assimilation with equivariance-preserving spatial transformers in a case study with era5

    Ashesh Chattopadhyay, Mustafa Mustafa, Pedram Hassanzadeh, Eviatar Bach, and Karthik Kashinath. Towards physically consistent data-driven weather forecasting: Integrating data assimilation with equivariance-preserving spatial transformers in a case study with era5. Geoscientific Model Development Discussions, pages 1--23, 2021

  18. [18]

    A machine learning-based global atmospheric forecast model

    Troy Arcomano, Istvan Szunyogh, Jaideep Pathak, Alexander Wikner, Brian R Hunt, and Edward Ott. A machine learning-based global atmospheric forecast model. Geophysical Research Letters, 47 0 (9): 0 e2020GL087776, 2020

  19. [19]

    Opportunities and challenges for machine learning in weather and climate modelling: hard, medium and soft ai

    Matthew Chantry, Hannah Christensen, Peter Dueben, and Tim Palmer. Opportunities and challenges for machine learning in weather and climate modelling: hard, medium and soft ai. Philosophical Transactions of the Royal Society A, 379 0 (2194): 0 20200083, 2021

  20. [20]

    Deep learning for post-processing ensemble weather forecasts

    Peter Gr \"o nquist, Chengyuan Yao, Tal Ben-Nun, Nikoli Dryden, Peter Dueben, Shigang Li, and Torsten Hoefler. Deep learning for post-processing ensemble weather forecasts. Philosophical Transactions of the Royal Society A, 379 0 (2194): 0 20200092, 2021

  21. [21]

    Data-driven medium-range weather prediction with a resnet pretrained on climate simulations: A new model for weatherbench

    Stephan Rasp and Nils Thuerey. Data-driven medium-range weather prediction with a resnet pretrained on climate simulations: A new model for weatherbench. Journal of Advances in Modeling Earth Systems, page e2020MS002405, 2021 b

  22. [22]

    Adaptive Fourier Neural Operators : Efficient token mixers for transformers

    John Guibas, Morteza Mardani, Zongyi Li, Andrew Tao, Anima Anandkumar, and Bryan Catanzaro. Adaptive Fourier Neural Operators : Efficient token mixers for transformers. International Conference on Representation Learning (to appear), April 2022

  23. [23]

    An image is worth 16x16 words: Transformers for image recognition at scale, 2021

    Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, Jakob Uszkoreit, and Neil Houlsby. An image is worth 16x16 words: Transformers for image recognition at scale, 2021

  24. [24]

    Fourier neural operator for parametric partial differential equations

    Zongyi Li, Nikola Kovachki, Kamyar Azizzadenesheli, Burigede Liu, Kaushik Bhattacharya, Andrew Stuart, and Anima Anandkumar. Fourier neural operator for parametric partial differential equations. In International Conference on Learning Representations (ICLR), 2021 a

  25. [25]

    Hans Hersbach, Bill Bell, Paul Berrisford, Shoji Hirahara, Andr \'a s Hor \'a nyi, Joaqu \'i n Mu \ n oz-Sabater, Julien Nicolas, Carole Peubey, Raluca Radu, Dinand Schepers, Adrian Simmons, Cornel Soci, Saleh Abdalla, Xavier Abellan, Gianpaolo Balsamo, Peter Bechtold, Gionata Biavati, Jean Bidlot, Massimo Bonavita, Giovanna De Chiara, Per Dahlgren, Dick ...

  26. [26]

    The ncep/ncar 40-year reanalysis project

    Eugenia Kalnay, Masao Kanamitsu, Robert Kistler, William Collins, Dennis Deaven, Lev Gandin, Mark Iredell, Suranjana Saha, Glenn White, John Woollen, et al. The ncep/ncar 40-year reanalysis project. Bulletin of the American meteorological Society, 77 0 (3): 0 437--472, 1996

  27. [27]

    Atmospheric modeling, data assimilation and predictability

    Eugenia Kalnay. Atmospheric modeling, data assimilation and predictability. Cambridge University Press, 2003

  28. [28]

    Beven II, R

    J.L. Beven II, R. Berg, and A. Hagen. Tropical cyclone report hurricane michael, April 2019

  29. [29]

    The ecmwf ensemble prediction system: Looking back (more than) 25 years and projecting forward 25 years

    Tim Palmer. The ecmwf ensemble prediction system: Looking back (more than) 25 years and projecting forward 25 years. Quarterly Journal of the Royal Meteorological Society, 145: 0 12--24, 2019

  30. [30]

    The ecmwf scalability programme: Progress and plans

    Peter Bauer, Tiago Quintino, Nils Wedi, Antonio Bonanni, Marcin Chrust, Willem Deconinck, Michail Diamantakis, Peter D \"u ben, Stephen English, Johannes Flemming, et al. The ecmwf scalability programme: Progress and plans. European Centre for Medium Range Weather Forecasts, 2020. doi:10.21957/gdit22ulm. URL https://www.ecmwf.int/node/19380

  31. [31]

    The ensemble kalman filter: Theoretical formulation and practical implementation

    Geir Evensen. The ensemble kalman filter: Theoretical formulation and practical implementation. Ocean dynamics, 53 0 (4): 0 343--367, 2003

  32. [32]

    Collins, and Caroline Muller

    Benjamin Fildier, William D. Collins, and Caroline Muller. Distortions of the rain distribution with warming, with and without self-aggregation. Journal of Advances in Modeling Earth Systems, 13 0 (2): 0 e2020MS002256, 2021. doi:https://doi.org/10.1029/2020MS002256. URL https://agupubs.onlinelibrary.wiley.com/doi/abs/10.1029/2020MS002256. e2020MS002256 20...

  33. [33]

    Deep spatial transformers for autoregressive data-driven forecasting of geophysical turbulence

    Ashesh Chattopadhyay, Mustafa Mustafa, Pedram Hassanzadeh, and Karthik Kashinath. Deep spatial transformers for autoregressive data-driven forecasting of geophysical turbulence. In Proceedings of the 10th International Conference on Climate Informatics, pages 106--112, 2020 b

  34. [34]

    Zero: Memory optimizations toward training trillion parameter models

    Samyam Rajbhandari, Jeff Rasley, Olatunji Ruwase, and Yuxiong He. Zero: Memory optimizations toward training trillion parameter models. In SC20: International Conference for High Performance Computing, Networking, Storage and Analysis, pages 1--16. IEEE, 2020

  35. [35]

    URL https://www.microsoft.com/en-us/research/blog/deepspeed-accelerating-large-scale-model-inference-and-training-via-system-optimizatio/ns-and-compression/

    Deep S peed: Accelerating large-scale model inference and training via system optimizations and compression, 2021. URL https://www.microsoft.com/en-us/research/blog/deepspeed-accelerating-large-scale-model-inference-and-training-via-system-optimizatio/ns-and-compression/

  36. [36]

    Physics-informed machine learning: case studies for weather and climate modelling

    K Kashinath, M Mustafa, A Albert, JL Wu, C Jiang, S Esmaeilzadeh, K Azizzadenesheli, R Wang, A Chattopadhyay, A Singh, et al. Physics-informed machine learning: case studies for weather and climate modelling. Philosophical Transactions of the Royal Society A, 379 0 (2194): 0 20200093, 2021

  37. [37]

    Physics-informed neural operator for learning partial differential equations, 2021 b

    Zongyi Li, Hongkai Zheng, Nikola Kovachki, David Jin, Haoxuan Chen, Burigede Liu, Kamyar Azizzadenesheli, and Anima Anandkumar. Physics-informed neural operator for learning partial differential equations, 2021 b

  38. [38]

    A hybrid approach to atmospheric modeling that combines machine learning with a physics-based numerical model

    Troy Arcomano, Istvan Szunyogh, Alexander Wikner, Jaideep Pathak, Brian R Hunt, and Edward Ott. A hybrid approach to atmospheric modeling that combines machine learning with a physics-based numerical model. Journal of Advances in Modeling Earth Systems, 2021