pith. machine review for the scientific record. sign in

arxiv: 2605.10297 · v1 · submitted 2026-05-11 · 💻 cs.CE

Recognition: no theorem link

QuantWeather: Quantile-Aware Probabilistic Forecasting for Subseasonal Precipitation

Authors on Pith no claims yet

Pith reviewed 2026-05-12 04:34 UTC · model grok-4.3

classification 💻 cs.CE
keywords probabilistic forecastingsubseasonal precipitationquantile regressiondual-head neural networkend-to-end learninguncertainty estimationweather prediction
0
0 comments X

The pith

A dual-head neural network produces reliable probabilistic subseasonal precipitation forecasts without post-hoc calibration.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper introduces QuantWeather, a framework that directly models predictive distributions for precipitation forecasts at subseasonal timescales. It uses a dual-head architecture where one head predicts deterministic values and the other handles probabilistic outputs through quantile awareness. These heads are trained with separate loss functions but optimized together in an end-to-end manner. This setup aims to generate well-calibrated uncertainty estimates from a single model run, avoiding the need for expensive ensemble simulations and subsequent calibration steps that rely on historical reforecasts. If successful, it would make reliable uncertainty information more accessible for applications like agriculture and disaster preparedness while lowering the computational burden during forecasting.

Core claim

QuantWeather is an end-to-end probabilistic forecasting framework featuring a dual-head design. The probabilistic head and deterministic head are supervised using separate objectives and jointly optimized. The model supports stochastic sampling, which allows it to generate probabilistic outputs even from a single forward pass, with the option to aggregate multiple samples if desired. Experiments indicate that this approach achieves superior probabilistic forecasting skill compared to existing methods while significantly cutting down on inference-time computation and storage requirements.

What carries the argument

The dual-head architecture with separate supervision objectives for deterministic and probabilistic predictions, enabling joint optimization and stochastic sampling for direct probabilistic outputs.

Load-bearing premise

That jointly supervising the deterministic and probabilistic heads with separate objectives produces well-calibrated predictive distributions directly from the model, eliminating the need for post-hoc calibration on reforecast datasets.

What would settle it

A direct comparison showing that QuantWeather's predictive distributions match observed precipitation frequencies better than uncalibrated ensemble forecasts, without any post-processing, using standard metrics like the Continuous Ranked Probability Score on held-out data.

Figures

Figures reproduced from arXiv: 2605.10297 by Hao Li, Lei Chen, Xiaohui Zhong, Xinyu Su.

Figure 1
Figure 1. Figure 1: Date-conditioned observed and model￾derived (FuXi-S2S) q80 climatological thresholds at Week 3. Each sample corresponds to a grid cell and an initialization date. The systematic deviation below the diagonal indicates distribution mismatch in raw forecasts. Forecasting total precipitation at the subseasonal scale, i.e., at lead times of roughly two to six weeks, is important for disaster preparedness, water… view at source ↗
Figure 2
Figure 2. Figure 2: Architecture of QuantWeather framework. QuantWeather adopts a dual-head design, consisting of a regression head that produces regression forecast Yˆ t+1 for autoregressive rollout, and a probabilistic head that produces classification probabilities Qˆ t+1 . The Perturb Module models stochastic variability. The Perturb Module models stochastic variability via two encoders. The posterior encoder Q takes {Xt … view at source ↗
Figure 3
Figure 3. Figure 3: Average RPSS and BSS without latitude weighting for total precipitation (TP) at forecast lead times of weeks 5 and 6, evaluated using all testing data from 2022. Values closer to 1 indicate better skill for both metrics. The first three columns show ECMWF-S2S, FuXi-S2S, and QuantWeather, and the fourth column shows the difference between QuantWeather and the best baseline, FuXi-S2S. Red contour lines mark … view at source ↗
Figure 4
Figure 4. Figure 4: Comparison of latitude-weighted RPSS and BSS for TP forecasts from ECMWF-S2S, FuXi-S2S, and QuantWeather, evaluated using all testing data from 2022. Results are averaged over global, land, and sea regions, corresponding to the three columns. The two rows show RPSS and BSS, respectively, across forecast lead times from week 3 to week 6. Pale bars indicate cases where QuantWeather does not show a statistica… view at source ↗
Figure 5
Figure 5. Figure 5: Average RPSS without latitude weighting for total precipitation (TP) at forecast lead times of week 3 and week4, evaluated using all testing data from 2022. The first three columns show the RPSS of the baselines and QuantWeather, and the fourth column shows the RPSS difference between QuantWeather and the best baseline, FuXi-S2S. Red contour lines mark positive RPSS in the first three columns and positive … view at source ↗
Figure 6
Figure 6. Figure 6: Comparison of latitude-weighted RPSS and BSS for TP forecasts from EC-S2S, FuXi-S2S, and QuantWeatherwith same number of ensemble member (M=8), evaluated using all testing data from 2022. Results are averaged over global, land, and sea regions, corresponding to the three columns. The two rows show RPSS and BSS, respectively, across forecast lead times from week 3 to week 6. Pale bars indicate cases where Q… view at source ↗
Figure 7
Figure 7. Figure 7: Parameter study. Varying number of ensemble member from 1 to 32. Report latitude-averaged RPSS and BSS total precipitation (TP) at forecast lead times of week 6, evaluated using all testing data from 2022. D.6 Parameter Study We study the impact of the number of stochastic members used during inference, varying it over {1, 2, 4, 8, 16, 32} [PITH_FULL_IMAGE:figures/full_fig_p019_7.png] view at source ↗
Figure 8
Figure 8. Figure 8: Comparison of anomaly-field RMSE, TCC, and ACC for regression forecasts using different numbers of inference members. The results are reported from Week 3 to Week 6 over global and land regions. Lower RMSE is better, while higher TCC and ACC indicate better performance. 21 [PITH_FULL_IMAGE:figures/full_fig_p021_8.png] view at source ↗
read the original abstract

Subseasonal precipitation forecasting is inherently uncertain due to chaotic atmospheric dynamics, making reliable uncertainty estimation essential for real-world applications. Existing approaches typically represent uncertainty through ensemble forecasts rather than directly modeling predictive distributions. However, due to systematic model biases, raw ensemble outputs are often not well calibrated and cannot be directly interpreted as reliable uncertainty estimates. As a result, operational systems rely on post-hoc calibration based on reforecast datasets, which are computationally expensive to generate and maintain. To address these limitations, we propose QuantWeather, an end-to-end probabilistic forecasting framework with a dual-head design. The probabilistic and deterministic heads are supervised with separate objectives and optimized jointly. The framework further supports stochastic sampling, enabling probabilistic outputs even with a single stochastic forward pass and allowing optional multi-sample aggregation. Extensive experiments show that QuantWeather demonstrates superior probabilistic forecasting skill while substantially reducing inference-time computational and storage costs.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

3 major / 2 minor

Summary. The paper proposes QuantWeather, a dual-head neural architecture for subseasonal precipitation forecasting that jointly optimizes a deterministic head and a quantile-aware probabilistic head under separate objectives. It claims this yields well-calibrated predictive distributions directly from a single forward pass (with optional stochastic sampling), eliminating the need for expensive post-hoc calibration on reforecast datasets while delivering superior probabilistic skill and lower inference-time compute/storage costs compared to ensemble-based approaches.

Significance. If the empirical claims are substantiated, the framework could meaningfully reduce operational barriers in subseasonal forecasting by removing reliance on large reforecast archives and enabling efficient, single-pass probabilistic outputs. The dual-head design with explicit quantile supervision is a targeted response to the calibration problem in chaotic atmospheric models.

major comments (3)
  1. Abstract: the central claim of 'superior probabilistic forecasting skill' and 'substantially reducing inference-time computational and storage costs' is stated without any quantitative metrics (e.g., CRPS, Brier scores, reliability diagrams, or wall-clock comparisons), baseline names, or data-split details, making it impossible to assess whether the dual-head design actually outperforms post-hoc calibrated ensembles.
  2. Method section (dual-head supervision): the assertion that separate deterministic and quantile objectives produce well-calibrated distributions directly, without post-hoc recalibration, lacks supporting analysis or ablation; in subseasonal regimes where ensemble spread dominates uncertainty, joint supervision alone does not guarantee that the learned quantiles match the true conditional distribution, and no diagnostic (PIT, coverage plots) is referenced to isolate this effect.
  3. Experiments: no evidence is supplied that raw QuantWeather outputs were compared against post-calibrated baselines or that calibration diagnostics were computed on held-out reforecast periods, leaving the 'no post-hoc calibration required' claim unverified and load-bearing for the cost-reduction argument.
minor comments (2)
  1. Abstract and introduction: the phrase 'stochastic sampling, enabling probabilistic outputs even with a single stochastic forward pass' is ambiguous; clarify whether this refers to dropout at inference, learned noise injection, or another mechanism.
  2. Notation: quantile levels and the exact form of the probabilistic loss are not defined in the provided abstract; ensure they appear explicitly in the methods with reference to standard pinball or quantile loss formulations.

Simulated Author's Rebuttal

3 responses · 0 unresolved

We thank the referee for the constructive and detailed comments. We address each major comment point-by-point below, with clear indications of planned revisions to the manuscript.

read point-by-point responses
  1. Referee: Abstract: the central claim of 'superior probabilistic forecasting skill' and 'substantially reducing inference-time computational and storage costs' is stated without any quantitative metrics (e.g., CRPS, Brier scores, reliability diagrams, or wall-clock comparisons), baseline names, or data-split details, making it impossible to assess whether the dual-head design actually outperforms post-hoc calibrated ensembles.

    Authors: We agree that the abstract is high-level and omits specific numbers. The manuscript's Experiments section (Section 4) reports quantitative results including CRPS and Brier score improvements over named baselines (raw and post-hoc calibrated ensembles), wall-clock timings, and data-split details (e.g., training on 2000-2015, validation 2016-2018, test 2019-2022). To address the concern, we will revise the abstract to incorporate concise quantitative highlights such as 'X% lower CRPS than post-calibrated ensembles with Y% reduced inference cost' while remaining within length limits. revision: yes

  2. Referee: Method section (dual-head supervision): the assertion that separate deterministic and quantile objectives produce well-calibrated distributions directly, without post-hoc recalibration, lacks supporting analysis or ablation; in subseasonal regimes where ensemble spread dominates uncertainty, joint supervision alone does not guarantee that the learned quantiles match the true conditional distribution, and no diagnostic (PIT, coverage plots) is referenced to isolate this effect.

    Authors: The dual-head architecture uses separate loss terms (MSE for the deterministic head and quantile loss for the probabilistic head) to encourage both point accuracy and distributional calibration in a single model. We acknowledge the value of explicit verification. In the revision we will add an ablation comparing dual-head versus single-head variants and include PIT histograms plus empirical coverage plots (at 10%, 50%, 90% quantiles) computed on held-out data to isolate the calibration effect of the joint supervision. revision: yes

  3. Referee: Experiments: no evidence is supplied that raw QuantWeather outputs were compared against post-calibrated baselines or that calibration diagnostics were computed on held-out reforecast periods, leaving the 'no post-hoc calibration required' claim unverified and load-bearing for the cost-reduction argument.

    Authors: The Experiments section does present direct comparisons of raw QuantWeather quantile outputs against both raw ensemble forecasts and post-hoc calibrated ensemble baselines (using standard reforecast-based methods), with skill scores and reliability diagrams shown for held-out reforecast periods. The cost savings are quantified via single-pass inference versus ensemble generation plus calibration overhead. To make the comparison more explicit and address the concern, we will add a dedicated table and text clarifying that all reported QuantWeather results use raw outputs without any post-hoc step, while still outperforming the calibrated baselines. revision: partial

Circularity Check

0 steps flagged

No circularity; empirical framework evaluated against external observations

full rationale

The paper introduces QuantWeather as an end-to-end trainable dual-head neural architecture whose probabilistic outputs are produced by joint optimization of separate deterministic and quantile objectives, then directly compared to held-out observational data for skill assessment. No equations, uniqueness theorems, or predictions are presented that reduce by construction to fitted inputs or self-citation chains; the central claims rest on experimental benchmarks rather than definitional equivalence or imported ansatzes. The framework is therefore self-contained against external verification.

Axiom & Free-Parameter Ledger

0 free parameters · 0 axioms · 0 invented entities

Abstract-only review provides no explicit free parameters, axioms, or invented entities. Standard neural-network assumptions (universal approximation, gradient-based optimization) are implicit but not stated.

pith-pipeline@v0.9.0 · 5451 in / 1129 out tokens · 34415 ms · 2026-05-12T04:34:47.999932+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

36 extracted references · 36 canonical work pages · 1 internal anchor

  1. [1]

    Potential Applications of Subseasonal-to-seasonal (S2S) Predictions.Meteorological applications, 24:315–325, 2017

    Christopher J White, Henrik Carlsen, Andrew W Robertson, Richard JT Klein, Jeffrey K Lazo, Arun Kumar, Frederic Vitart, Erin Coughlan de Perez, Andrea J Ray, Virginia Murray, et al. Potential Applications of Subseasonal-to-seasonal (S2S) Predictions.Meteorological applications, 24:315–325, 2017

  2. [2]

    The Subseasonal Experi- ment (SubX): A Multimodel Subseasonal Prediction Experiment.Bulletin of the American Meteorological Society, 100:2043–2060, 2019

    Kathy Pegion, Ben P Kirtman, Emily Becker, Dan C Collins, Emerson LaJoie, Robert Burgman, Ray Bell, Timothy DelSole, Dughong Min, Yuejian Zhu, et al. The Subseasonal Experi- ment (SubX): A Multimodel Subseasonal Prediction Experiment.Bulletin of the American Meteorological Society, 100:2043–2060, 2019

  3. [3]

    Sub- seasonal Forecasts of Demand and Wind Power and Solar Power Generation for 28 European Countries.Earth System Science Data, 13:2259–2274, 2021

    Hannah C Bloomfield, David J Brayshaw, Paula LM Gonzalez, and Andrew Charlton-Perez. Sub- seasonal Forecasts of Demand and Wind Power and Solar Power Generation for 28 European Countries.Earth System Science Data, 13:2259–2274, 2021

  4. [4]

    Advances in the Application and Utility of Subseasonal-to-seasonal Predictions.Bulletin of the American Meteorological Society, 103:E1448–E1472, 2022

    Christopher J White, Daniela IV Domeisen, Nachiketa Acharya, Elijah A Adefisan, Michael L Anderson, Stella Aura, Ahmed A Balogun, Douglas Bertram, Sonia Bluhm, David J Brayshaw, et al. Advances in the Application and Utility of Subseasonal-to-seasonal Predictions.Bulletin of the American Meteorological Society, 103:E1448–E1472, 2022

  5. [5]

    Advances in the subseasonal prediction of extreme events: Relevant case studies across the globe.Bulletin of the American Meteorological Society, 103:E1473–E1501, 2022

    Daniela IV Domeisen, Christopher J White, Hilla Afargan-Gerstman, Ángel G Muñoz, Matthew A Janiga, Frédéric Vitart, C Ole Wulff, Salomé Antoine, Constantin Ardilouze, Lauriane Batté, et al. Advances in the subseasonal prediction of extreme events: Relevant case studies across the globe.Bulletin of the American Meteorological Society, 103:E1473–E1501, 2022

  6. [6]

    Forced and Free Variations of Weather and Climate.Journal of Atmospheric Sciences, 36:1367–1376, 1979

    Edward N Lorenz. Forced and Free Variations of Weather and Climate.Journal of Atmospheric Sciences, 36:1367–1376, 1979

  7. [7]

    Progress in Subseasonal to Seasonal Pre- diction through a Joint Weather and Climate Community Effort.Npj Climate and Atmospheric Science, 1:4, 2018

    Annarita Mariotti, Paolo M Ruti, and Michel Rixen. Progress in Subseasonal to Seasonal Pre- diction through a Joint Weather and Climate Community Effort.Npj Climate and Atmospheric Science, 1:4, 2018

  8. [8]

    A machine learning model that outperforms conventional global subseasonal forecast models.Nature Communications, 15(1):6425, 2024

    Lei Chen, Xiaohui Zhong, Hao Li, Jie Wu, Bo Lu, Deliang Chen, Shang-Ping Xie, Libo Wu, Qingchen Chao, Chensen Lin, et al. A machine learning model that outperforms conventional global subseasonal forecast models.Nature Communications, 15(1):6425, 2024

  9. [9]

    A comparison of the ECMWF, MSC, and NCEP global ensemble prediction systems

    Roberto Buizza, Peter L Houtekamer, Gérard Pellerin, Zoltan Toth, Yuejian Zhu, and Mozheng Wei. A comparison of the ECMWF, MSC, and NCEP global ensemble prediction systems. Monthly Weather Review, 133:1076–1097, 2005

  10. [10]

    Ensemble Forecasting.Journal of Computational Physics, 227:3515–3539, 2008

    Martin Leutbecher and Tim N Palmer. Ensemble Forecasting.Journal of Computational Physics, 227:3515–3539, 2008

  11. [11]

    Sub-seasonal Forecasting with a Large Ensemble of Deep-learning Weather Prediction Models.Journal of Advances in Modeling Earth Systems, 13:e2021MS002502, 2021

    Jonathan A Weyn, Dale R Durran, Rich Caruana, and Nathaniel Cresswell-Clay. Sub-seasonal Forecasting with a Large Ensemble of Deep-learning Weather Prediction Models.Journal of Advances in Modeling Earth Systems, 13:e2021MS002502, 2021

  12. [12]

    Ensemble size versus bias correction effects in subseasonal-to-seasonal (s2s) forecasts.Geoscience Letters, 10:37, 2023

    Ji-Young Han, Sang-Wook Kim, Chang-Hyun Park, and Seok-Woo Son. Ensemble size versus bias correction effects in subseasonal-to-seasonal (s2s) forecasts.Geoscience Letters, 10:37, 2023

  13. [13]

    Eviatar Bach, Venkat Krishnamurthy, Safa Mote, Jagadish Shukla, A Surjalal Sharma, Eugenia Kalnay, and Michael Ghil. Improved Subseasonal Prediction of South Asian Monsoon Rainfall using Data-driven Forecasts of Oscillatory Modes.Proceedings of the National Academy of Sciences, 121(15):e2312573121, 2024. 10

  14. [14]

    The ERA5 Global Reanalysis.Quarterly journal of the royal meteorological society, 146:1999–2049, 2020

    Hersbach, Hans and Bell, Bill and Berrisford, Paul and Hirahara, Shoji and Horányi, András and Muñoz-Sabater, Joaquín and Nicolas, Julien and Peubey, Carole and Radu, Raluca and Schepers, Dinand and others. The ERA5 Global Reanalysis.Quarterly journal of the royal meteorological society, 146:1999–2049, 2020

  15. [15]

    Fuxi-ens: A machine learning model for efficient and accurate ensemble weather prediction.Science Advances, 11:eadu2854, 2025

    Xiaohui Zhong, Lei Chen, Hao Li, Roberto Buizza, Jun Liu, Jie Feng, Zijian Zhu, Xu Fan, Kan Dai, Jing-jia Luo, et al. Fuxi-ens: A machine learning model for efficient and accurate ensemble weather prediction.Science Advances, 11:eadu2854, 2025

  16. [16]

    FourCastNet: A Global Data-driven High-resolution Weather Model using Adaptive Fourier Neural Operators

    Jaideep Pathak, Shashank Subramanian, Peter Harrington, Sanjeev Raja, Ashesh Chattopadhyay, Morteza Mardani, Thorsten Kurth, David Hall, Zongyi Li, Kamyar Azizzadenesheli, et al. Fourcastnet: A global data-driven high-resolution weather model using adaptive fourier neural operators.arXiv preprint arXiv:2202.11214, 2022

  17. [17]

    Accurate medium-range global weather forecasting with 3d neural networks.Nature, 619:533–538, 2023

    Kaifeng Bi, Lingxi Xie, Hengheng Zhang, Xin Chen, Xiaotao Gu, and Qi Tian. Accurate medium-range global weather forecasting with 3d neural networks.Nature, 619:533–538, 2023

  18. [18]

    Learning Skillful Medium-range Global Weather Forecasting.Science, 382(6677):1416–1421, 2023

    Remi Lam, Alvaro Sanchez-Gonzalez, Matthew Willson, Peter Wirnsberger, Meire Fortunato, Ferran Alet, Suman Ravuri, Timo Ewalds, Zach Eaton-Rosen, Weihua Hu, et al. Learning Skillful Medium-range Global Weather Forecasting.Science, 382(6677):1416–1421, 2023

  19. [19]

    Fengwu: Pushing the skillful global medium-range weather forecast beyond 10 days lead,

    Kang Chen, Tao Han, Junchao Gong, Lei Bai, Fenghua Ling, Jing-Jia Luo, Xi Chen, Leiming Ma, Tianning Zhang, Rui Su, et al. Fengwu: Pushing the skillful global medium-range weather forecast beyond 10 days lead.arXiv preprint arXiv:2304.02948, 2023

  20. [20]

    Fuxi: A cascade machine learning forecasting system for 15-day global weather forecast.npj climate and atmospheric science, 6(1):190, 2023

    Lei Chen, Xiaohui Zhong, Feng Zhang, Yuan Cheng, Yinghui Xu, Yuan Qi, and Hao Li. Fuxi: A cascade machine learning forecasting system for 15-day global weather forecast.npj climate and atmospheric science, 6(1):190, 2023

  21. [21]

    Neural general circulation models for weather and climate.Nature, 632(8027):1060–1066, 2024

    Dmitrii Kochkov, Janni Yuval, Ian Langmore, Peter Norgaard, Jamie Smith, Griffin Mooers, Milan Klöwer, James Lottes, Stephan Rasp, Peter Düben, et al. Neural general circulation models for weather and climate.Nature, 632(8027):1060–1066, 2024

  22. [22]

    FuXi-RTM: A Physics-Guided Pre- diction Framework with Radiative Transfer Modeling

    Qiusheng Huang, Xiaohui Zhong, Xu Fan, and Hao Li. FuXi-RTM: A Physics-Guided Pre- diction Framework with Radiative Transfer Modeling. InProceedings of the IEEE/CVF International Conference on Computer Vision, pages 8852–8862, 2025

  23. [23]

    Cirt: Global subseasonal-to-seasonal forecasting with geometry-inspired transformer.arXiv preprint arXiv:2502.19750, 2025

    Yang Liu, Zinan Zheng, Jiashun Cheng, Fugee Tsung, Deli Zhao, Yu Rong, and Jia Li. Cirt: Global subseasonal-to-seasonal forecasting with geometry-inspired transformer.arXiv preprint arXiv:2502.19750, 2025

  24. [24]

    Equivariant and Invariant Message Passing for Global Subseasonal-to-seasonal Forecasting

    Yang Liu, Zinan Zheng, Yu Rong, Deli Zhao, Hong Cheng, and Jia Li. Equivariant and Invariant Message Passing for Global Subseasonal-to-seasonal Forecasting. InProceedings of the 31st ACM SIGKDD Conference on Knowledge Discovery and Data Mining V . 2, pages 1879–1890, 2025

  25. [25]

    Probabilistic weather forecasting with machine learning.Nature, 637:84–90, 2025

    Ilan Price, Alvaro Sanchez-Gonzalez, Ferran Alet, Tom R Andersson, Andrew El-Kadi, Dominic Masters, Timo Ewalds, Jacklynn Stott, Shakir Mohamed, Peter Battaglia, et al. Probabilistic weather forecasting with machine learning.Nature, 637:84–90, 2025

  26. [26]

    Generative emula- tion of weather forecast ensembles with diffusion models.Science Advances, 10(13):eadk4489, 2024

    Lizao Li, Robert Carver, Ignacio Lopez-Gomez, Fei Sha, and John Anderson. Generative emula- tion of weather forecast ensembles with diffusion models.Science Advances, 10(13):eadk4489, 2024

  27. [27]

    Calibrated probabilistic forecasting using ensemble model output statistics and minimum crps estimation

    Tilmann Gneiting, Adrian E Raftery, Anton H Westveld III, and Tom Goldman. Calibrated probabilistic forecasting using ensemble model output statistics and minimum crps estimation. Monthly weather review, 133(5):1098–1118, 2005

  28. [28]

    Using bayesian model averaging to calibrate forecast ensembles.Monthly weather review, 133(5):1155– 1174, 2005

    Adrian E Raftery, Tilmann Gneiting, Fadoua Balabdaoui, and Michael Polakowski. Using bayesian model averaging to calibrate forecast ensembles.Monthly weather review, 133(5):1155– 1174, 2005. 11

  29. [29]

    Calibrated Ensemble Forecasts using Quantile Regression Forests and Ensemble Model Output Statistics.Monthly Weather Review, 144(6):2375–2393, 2016

    Maxime Taillardat, Olivier Mestre, Michaël Zamo, and Philippe Naveau. Calibrated Ensemble Forecasts using Quantile Regression Forests and Ensemble Model Output Statistics.Monthly Weather Review, 144(6):2375–2393, 2016

  30. [30]

    Neural Networks for Postprocessing Ensemble Weather Forecasts.Monthly Weather Review, 146:3885–3900, 2018

    Stephan Rasp and Sebastian Lerch. Neural Networks for Postprocessing Ensemble Weather Forecasts.Monthly Weather Review, 146:3885–3900, 2018

  31. [31]

    Ensemble Postprocessing using Quantile Function Regression based on Neural Networks and Bernstein Polynomials.Monthly Weather Review, 148:403–414, 2020

    John Bjørnar Bremnes. Ensemble Postprocessing using Quantile Function Regression based on Neural Networks and Bernstein Polynomials.Monthly Weather Review, 148:403–414, 2020

  32. [32]

    Using artificial neural networks for generating probabilistic subseasonal precipitation forecasts over california.Monthly Weather Review, 148(8):3489–3506, 2020

    Michael Scheuerer, Matthew B Switanek, Rochelle P Worsnop, and Thomas M Hamill. Using artificial neural networks for generating probabilistic subseasonal precipitation forecasts over california.Monthly Weather Review, 148(8):3489–3506, 2020

  33. [33]

    K., Espeholt, L., Heek, J., Dehghani, M., Oliver, A., Salimans, T., Agrawal, S., Hickey, J., and Kalchbrenner, N

    Casper Kaae Sønderby, Lasse Espeholt, Jonathan Heek, Mostafa Dehghani, Avital Oliver, Tim Salimans, Shreya Agrawal, Jason Hickey, and Nal Kalchbrenner. MetNet: A Neural Weather Model for Precipitation Forecasting.arXiv preprint arXiv:2003.12140, 2020

  34. [34]

    Deep learning for twelve hour precipitation forecasts.Nature communications, 13(1):5145, 2022

    Lasse Espeholt, Shreya Agrawal, Casper Sønderby, Manoj Kumar, Jonathan Heek, Carla Bromberg, Cenk Gazen, Rob Carver, Marcin Andrychowicz, Jason Hickey, et al. Deep learning for twelve hour precipitation forecasts.Nature communications, 13(1):5145, 2022

  35. [35]

    The ai weather quest: an international competition for sub-seasonal forecasting with ai.Machine Learning: Earth, 1:010701, 2025

    Olga Loegel, Joshua Talib, Frederic Vitart, Jörn Hoffmann, and Matthew Chantry. The ai weather quest: an international competition for sub-seasonal forecasting with ai.Machine Learning: Earth, 1:010701, 2025

  36. [36]

    A foundation model for the earth system.Nature, 641:1180–1187, 2025

    Cristian Bodnar, Wessel P Bruinsma, Ana Lucic, Megan Stanley, Anna Allen, Johannes Brand- stetter, Patrick Garvan, Maik Riechert, Jonathan A Weyn, Haiyu Dong, et al. A foundation model for the earth system.Nature, 641:1180–1187, 2025. 12 A Related Work A.1 Deterministic Weather Forecasting Deterministic forecasting has been one of the dominant paradigms i...