Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
173 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learning to forecast diagnostic parameters using pre-trained weather embedding (2312.00290v1)

Published 1 Dec 2023 in cs.LG

Abstract: Data-driven weather prediction (DDWP) models are increasingly becoming popular for weather forecasting. However, while operational weather forecasts predict a wide variety of weather variables, DDWPs currently forecast a specific set of key prognostic variables. Non-prognostic ("diagnostic") variables are sometimes modeled separately as dependent variables of the prognostic variables (c.f. FourCastNet), or by including the diagnostic variable as a target in the DDWP. However, the cost of training and deploying bespoke models for each diagnostic variable can increase dramatically with more diagnostic variables, and limit the operational use of such models. Likewise, retraining an entire DDWP each time a new diagnostic variable is added is also cost-prohibitive. We present an two-stage approach that allows new diagnostic variables to be added to an end-to-end DDWP model without the expensive retraining. In the first stage, we train an autoencoder that learns to embed prognostic variables into a latent space. In the second stage, the autoencoder is frozen and "downstream" models are trained to predict diagnostic variables using only the latent representations of prognostic variables as input. Our experiments indicate that models trained using the two-stage approach offer accuracy comparable to training bespoke models, while leading to significant reduction in resource utilization during training and inference. This approach allows for new "downstream" models to be developed as needed, without affecting existing models and thus reducing the friction in operationalizing new models.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (14)
  1. Accurate medium-range global weather forecasting with 3d neural networks. Nature, pages 1–6.
  2. Dynamical tests of a deep-learning weather prediction model. arXiv preprint arXiv:2309.10867.
  3. The ERA5 global reanalysis. Quarterly Journal of the Royal Meteorological Society, 146(730):1999–2049.
  4. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980.
  5. Graphcast: Learning skillful medium-range global weather forecasting.
  6. Neighborhood reconstructing autoencoders. Advances in Neural Information Processing Systems, 34:536–546.
  7. How do transformers learn topic structure: Towards a mechanistic understanding. arXiv preprint arXiv:2303.04245.
  8. W-mae: Pre-trained weather model with masked autoencoder for multi-variable weather forecasting. arXiv preprint arXiv:2304.08754.
  9. Climax: A foundation model for weather and climate. arXiv preprint arXiv:2301.10343.
  10. FourCastNet: A Global Data-driven High-resolution Weather Model using Adaptive Fourier Neural Operators.
  11. Verification against in-situ observations for data-driven weather prediction. arXiv preprint arXiv:2305.00048.
  12. Weatherbench 2: A benchmark for the next generation of data-driven global weather models. arXiv preprint arXiv:2308.15560.
  13. Visualizing and understanding convolutional networks. In Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part I 13, pages 818–833. Springer.
  14. Understanding the robustness in vision transformers. In International Conference on Machine Learning, pages 27378–27394. PMLR.
Citations (1)

Summary

We haven't generated a summary for this paper yet.