Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
156 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

CGNSDE: Conditional Gaussian Neural Stochastic Differential Equation for Modeling Complex Systems and Data Assimilation (2404.06749v1)

Published 10 Apr 2024 in cs.LG

Abstract: A new knowledge-based and machine learning hybrid modeling approach, called conditional Gaussian neural stochastic differential equation (CGNSDE), is developed to facilitate modeling complex dynamical systems and implementing analytic formulae of the associated data assimilation (DA). In contrast to the standard neural network predictive models, the CGNSDE is designed to effectively tackle both forward prediction tasks and inverse state estimation problems. The CGNSDE starts by exploiting a systematic causal inference via information theory to build a simple knowledge-based nonlinear model that nevertheless captures as much explainable physics as possible. Then, neural networks are supplemented to the knowledge-based model in a specific way, which not only characterizes the remaining features that are challenging to model with simple forms but also advances the use of analytic formulae to efficiently compute the nonlinear DA solution. These analytic formulae are used as an additional computationally affordable loss to train the neural networks that directly improve the DA accuracy. This DA loss function promotes the CGNSDE to capture the interactions between state variables and thus advances its modeling skills. With the DA loss, the CGNSDE is more capable of estimating extreme events and quantifying the associated uncertainty. Furthermore, crucial physical properties in many complex systems, such as the translate-invariant local dependence of state variables, can significantly simplify the neural network structures and facilitate the CGNSDE to be applied to high-dimensional systems. Numerical experiments based on chaotic systems with intermittency and strong non-Gaussian features indicate that the CGNSDE outperforms knowledge-based regression models, and the DA loss further enhances the modeling skills of the CGNSDE.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (141)
  1. Jürgen Jost. Dynamical systems: examples of complex behaviour. Springer Science & Business Media, 2005.
  2. Turbulence: the legacy of AN Kolmogorov. Cambridge university press, 1995.
  3. Nonlinear dynamics and statistical theories for basic geophysical flows. Cambridge University Press, 2006.
  4. Geoffrey K Vallis. Atmospheric and oceanic fluid dynamics. Cambridge University Press, 2017.
  5. Ralph H Abraham. Complex dynamical systems. In Mathematical modelling in science and technology, pages 82–86. Elsevier, 1984.
  6. Rick Salmon. Lectures on geophysical fluid dynamics. Oxford University Press, 1998.
  7. Introduction to dynamical systems. Cambridge university press, 2002.
  8. Steven H Strogatz. Nonlinear dynamics and chaos with student solutions manual: With applications to physics, biology, chemistry, and engineering. CRC press, 2018.
  9. Introduction to applied nonlinear dynamical systems and chaos, volume 2. Springer, 2003.
  10. Nan Chen. Stochastic Methods for Modeling and Predicting Complex Dynamical Systems: Uncertainty Quantification, State Estimation, and Reduced-Order Models. Springer Nature, 2023.
  11. Henk A Dijkstra. Nonlinear climate dynamics. Cambridge University Press, 2013.
  12. TN Palmer. A nonlinear dynamical perspective on climate change. Weather, 48(10):314–326, 1993.
  13. Extreme events: Mechanisms and prediction. Applied Mechanics Reviews, 71(5), 2019.
  14. Attribution of climate extreme events. Nature Climate Change, 5(8):725–730, 2015.
  15. HK Moffatt. Extreme events in turbulent flow. Journal of Fluid Mechanics, 914:F1, 2021.
  16. Andrew Majda. Introduction to PDEs and Waves for the Atmosphere and Ocean, volume 9. American Mathematical Soc., 2003.
  17. Intermittency and the Lorenz model. Physics Letters A, 75(1-2):1–2, 1979.
  18. Eugenia Kalnay. Atmospheric modeling, data assimilation and predictability. Cambridge university press, 2003.
  19. Boris Khattatov William Lahoz and Richard Menard. Data assimilation. Springer, 2010.
  20. Filtering complex turbulent systems. Cambridge University Press, 2012.
  21. Geir Evensen et al. Data assimilation: the ensemble Kalman filter, volume 2. Springer, 2009.
  22. Data assimilation. Cham, Switzerland: Springer, 214, 2015.
  23. Climate science and the uncertainty monster. Bulletin of the American Meteorological Society, 92(12):1667–1682, 2011.
  24. Timothy DelSole. Predictability and information theory. Part I: Measures of predictability. Journal of the atmospheric sciences, 61(20):2425–2440, 2004.
  25. Paul N Edwards. Global climate science, uncertainty and politics: Data-laden models, model-filtered data. Science as culture, 8(4):437–472, 1999.
  26. Lessons in uncertainty quantification for turbulent dynamical systems. Discrete and Continuous Dynamical Systems, 32(9):3133, 2012.
  27. Model error, information barriers, state estimation and prediction in complex multiscale systems. Entropy, 20(9):644, 2018.
  28. Nicolaas G Van Kampen. Stochastic differential equations. Physics reports, 24(3):171–228, 1976.
  29. Ludwig Arnold. Stochastic differential equations. New York, 2, 1974.
  30. Stochastic differential equations. In Numerical solution of stochastic differential equations, pages 103–160. Springer, 1992.
  31. Stochastic differential equations. Springer, 2005.
  32. Andrew J Majda and Di Qi. Strategies for reduced-order models for predicting the statistical responses and uncertainty quantification in complex turbulent dynamical systems. SIAM Review, 60(3):491–549, 2018.
  33. The gnat method for nonlinear model reduction: effective implementation and application to computational fluid dynamics and turbulent flows. Journal of Computational Physics, 242:623–647, 2013.
  34. Reduced-order modelling for flow control, volume 528. Springer Science & Business Media, 2011.
  35. Modal analysis of fluid flows: Applications and outlook. AIAA journal, 58(3):998–1022, 2020.
  36. Data-driven filtered reduced order modeling of fluid flows. SIAM Journal on Scientific Computing, 40(3):B834–B857, 2018.
  37. Physics-informed machine learning approach for reconstructing Reynolds stress modeling discrepancies based on DNS data. Physical Review Fluids, 2(3):034603, 2017.
  38. Physics-informed machine learning approach for augmenting turbulence models: A comprehensive framework. Physical Review Fluids, 3(7):074602, 2018.
  39. Stochastic parameterization: Toward a new view of weather and climate models. Bulletin of the American Meteorological Society, 98(3):565–588, 2017.
  40. Toward a stochastic parameterization of ocean mesoscale eddies. Ocean Modelling, 79:1–20, 2014.
  41. Andrew Dawson and TN Palmer. Simulating weather regimes: Impact of model resolution and stochastic parameterization. Climate Dynamics, 44:2177–2193, 2015.
  42. Learning stochastic closures using ensemble Kalman inversion. Transactions of Mathematics and Its Applications, 5(1):tnab003, 2021.
  43. On closures for reduced order models–A spectrum of first-principle to machine-learned avenues. Physics of Fluids, 33(9):091301, 2021.
  44. Data-adaptive harmonic spectra and multilayer Stuart-Landau models. Chaos: An Interdisciplinary Journal of Nonlinear Science, 27(9):093110, 2017.
  45. Data-driven model reduction, wiener projections, and the koopman-mori-zwanzig formalism. Journal of Computational Physics, 424:109864, 2021.
  46. Data-driven variational multiscale reduced order models. Computer Methods in Applied Mechanics and Engineering, 373:113470, 2021.
  47. Dynamic data-driven reduced-order models. Computer Methods in Applied Mechanics and Engineering, 291:21–41, 2015.
  48. Data-driven pod-galerkin reduced order model for turbulent flows. Journal of Computational Physics, 416:109513, 2020.
  49. Data-driven model predictive control using random forests for building energy optimization and climate control. Applied energy, 226:1252–1272, 2018.
  50. Operator learning for continuous spatial-temporal model with gradient-based and derivative-free optimization methods. arXiv preprint arXiv:2311.11798, 2023.
  51. Machine learning with data assimilation and uncertainty quantification for dynamical systems: a review. IEEE/CAA Journal of Automatica Sinica, 10(6):1361–1387, 2023.
  52. Introduction to focus issue: When machine learning meets complex systems: Networks, chaos, and nonlinear dynamics. Chaos: An Interdisciplinary Journal of Nonlinear Science, 30(6), 2020.
  53. Data-driven science and engineering: Machine learning, dynamical systems, and control. Cambridge University Press, 2019.
  54. Lift & learn: Physics-informed machine learning for large-scale nonlinear dynamical systems. Physica D: Nonlinear Phenomena, 406:132401, 2020.
  55. Turbulence modeling in the age of data. Annual review of fluid mechanics, 51:357–377, 2019.
  56. Machine learning for fluid mechanics. Annual review of fluid mechanics, 52:477–508, 2020.
  57. Learning about structural errors in models of complex dynamical systems. arXiv preprint arXiv:2401.00035, 2023.
  58. Combining data assimilation and machine learning to emulate a dynamical model from sparse and noisy observations: A case study with the lorenz 96 model. Journal of Computational Science, 44:101171, 2020.
  59. Deep data assimilation: integrating deep learning with data assimilation. Applied Sciences, 11(3):1114, 2021.
  60. Data learning: Integrating data assimilation and machine learning. Journal of Computational Science, 58:101525, 2022.
  61. Combining machine learning and data assimilation to forecast dynamical systems from noisy partial observations. Chaos: An Interdisciplinary Journal of Nonlinear Science, 31(10):101103, 2021.
  62. Deep learning-enhanced ensemble-based data assimilation for high-dimensional nonlinear dynamical systems. Journal of Computational Physics, 477:111918, 2023.
  63. Machine learning methods for data assimilation. Computational Intelligence in Architecturing Complex Engineering Systems, pages 105–112, 2010.
  64. Auto-encoder based dimensionality reduction. Neurocomputing, 184:232–242, 2016.
  65. Linearly recurrent autoencoder networks for learning dynamics. SIAM Journal on Applied Dynamical Systems, 18(1):558–593, 2019.
  66. Learning koopman invariant subspaces for dynamic mode decomposition. Advances in neural information processing systems, 30, 2017.
  67. Machine learning for stochastic parameterization: Generative adversarial networks in the lorenz’96 model. Journal of Advances in Modeling Earth Systems, 12(3):e2019MS001896, 2020.
  68. Using data assimilation to train a hybrid forecast system that combines machine-learning and knowledge-based components. Chaos: An Interdisciplinary Journal of Nonlinear Science, 31(5), 2021.
  69. Deep learning to represent subgrid processes in climate models. Proceedings of the National Academy of Sciences, 115(39):9684–9689, 2018.
  70. Machine learning for model error inference and correction. Journal of Advances in Modeling Earth Systems, 12(12):e2020MS002232, 2020.
  71. A comparison of combined data assimilation and machine learning methods for offline and online model error correction. Journal of computational science, 55:101468, 2021.
  72. State, global, and local parameter estimation using local ensemble kalman filters: Applications to online machine learning of chaotic dynamics. Quarterly Journal of the Royal Meteorological Society, 148(746):2167–2193, 2022.
  73. Online model error correction with neural networks in the incremental 4d-var framework. Journal of Advances in Modeling Earth Systems, 15(9):e2022MS003474, 2023.
  74. Online learning of both state and dynamics using ensemble kalman filters. arXiv preprint arXiv:2006.03859, 2020.
  75. Observation error covariance specification in dynamical systems for data assimilation using recurrent neural networks. Neural Computing and Applications, 34(16):13149–13167, 2022.
  76. Cello: A fast algorithm for covariance estimation. In 2013 IEEE International Conference on Robotics and Automation, pages 3160–3167. IEEE, 2013.
  77. Deep inference for covariance estimation: Learning gaussian noise models for state estimation. In 2018 IEEE International Conference on Robotics and Automation (ICRA), pages 1436–1443. IEEE, 2018.
  78. Kalmannet: Neural network aided kalman filtering for partially known dynamics. IEEE Transactions on Signal Processing, 70:1532–1547, 2022.
  79. Dan–an optimal data assimilation framework based on machine learning recurrent networks. arXiv preprint arXiv:2010.09694, 2020.
  80. Neural network based kalman filters for the spatio-temporal interpolation of satellite-derived sea surface temperature. Remote Sensing, 10(12):1864, 2018.
  81. A deep learning approach to spatiotemporal sea surface height interpolation and estimation of deep currents in geostrophic ocean turbulence. Journal of Advances in Modeling Earth Systems, 13(1):e2019MS001965, 2021.
  82. Combining stochastic parameterized reduced-order models with machine learning for data assimilation and uncertainty quantification with partial observations. Journal of Advances in Modeling Earth Systems, 15(10):e2022MS003597, 2023.
  83. Data assimilation networks. Journal of Advances in Modeling Earth Systems, 15(4):e2022MS003353, 2023.
  84. Autodifferentiable ensemble Kalman filters. SIAM Journal on Mathematics of Data Science, 4(2):801–833, 2022.
  85. Conditional Gaussian systems for multiscale nonlinear stochastic systems: Prediction, state estimation and uncertainty quantification. Entropy, 20(7):509, 2018.
  86. Conditional Gaussian nonlinear system: A fast preconditioner and a cheap surrogate model for complex nonlinear systems. Chaos: An Interdisciplinary Journal of Nonlinear Science, 32(5):053122, 2022.
  87. Filtering nonlinear turbulent dynamical systems through conditional gaussian statistics. Monthly Weather Review, 144(12):4885–4917, 2016.
  88. Statistics of random processes II: Applications, volume 6. Springer Science & Business Media, 2013.
  89. Predicting the cloud patterns of the Madden-Julian oscillation through a low-order nonlinear stochastic model. Geophysical Research Letters, 41(15):5612–5619, 2014.
  90. An efficient and statistically accurate Lagrangian data assimilation algorithm with applications to discrete element sea ice models. Journal of Computational Physics, 455:111000, 2022.
  91. Information barriers for noisy Lagrangian tracers in filtering random incompressible flows. Nonlinearity, 27(9):2133, 2014.
  92. Model error in filtering random compressible flows utilizing noisy Lagrangian tracers. Monthly Weather Review, 144(11):4037–4061, 2016.
  93. Lagrangian descriptors with uncertainty. arXiv preprint arXiv:2307.04006, 2023.
  94. Filtering the stochastic skeleton model for the Madden–Julian oscillation. Monthly Weather Review, 144(2):501–527, 2016.
  95. Predicting monsoon intraseasonal precipitation using a low-order nonlinear stochastic model. Journal of Climate, 31(11):4403–4427, 2018.
  96. Beating the curse of dimension with accurate statistics for the Fokker–Planck equation in complex turbulent systems. Proceedings of the National Academy of Sciences, 114(49):12864–12869, 2017.
  97. Efficient statistically accurate algorithms for the Fokker–Planck equation in large dimensions. Journal of Computational Physics, 354:242–268, 2018.
  98. New results in linear filtering and prediction theory. Journal of Fluids Engineering, 83:95–108, 1961.
  99. Dynamic stochastic superresolution of sparsely observed turbulent systems. Journal of Computational Physics, 241:333–363, 2013.
  100. New methods for estimating ocean eddy heat transport using satellite altimetry. Monthly Weather Review, 140(5):1703–1722, 2012.
  101. Mathematical test models for superparametrization in anisotropic turbulence. Proceedings of the National Academy of Sciences, 106(14):5470–5474, 2009.
  102. Stochastic superparameterization in a one-dimensional model for wave turbulence. Communications in Mathematical Sciences, 12(3):509–525, 2014.
  103. New perspectives on superparameterization for geophysical turbulence. Journal of Computational Physics, 271:60–77, 2014.
  104. Blended particle filters for large-dimensional chaotic dynamical systems. Proceedings of the National Academy of Sciences, 111(21):7511–7516, 2014.
  105. Lemda: A lagrangian-eulerian multiscale data assimilation framework. arXiv preprint arXiv:2401.18048, 2024.
  106. Models for stochastic climate prediction. Proceedings of the National Academy of Sciences, 96(26):14687–14691, 1999.
  107. A mathematical framework for stochastic climate models. Communications on Pure and Applied Mathematics: A Journal Issued by the Courant Institute of Mathematical Sciences, 54(8):891–974, 2001.
  108. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016.
  109. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE international conference on computer vision, pages 1026–1034, 2015.
  110. How entropic regression beats the outliers problem in nonlinear system identification. Chaos: An Interdisciplinary Journal of Nonlinear Science, 30(1), 2020.
  111. Erfit: Entropic regression fit MATLAB package, for data-driven system identification of underlying dynamic equations. arXiv preprint arXiv:2010.02411, 2020.
  112. Causation entropy method for covariate selection in dynamic models. In 2021 American Control Conference (ACC), pages 2842–2847. IEEE, 2021.
  113. A causality-based learning approach for discovering the underlying dynamics of complex systems from partial observations with stochastic parameterization. Physica D: Nonlinear Phenomena, 449:133743, 2023.
  114. CEBoosting: Online sparse identification of dynamical systems with regime switching by causation entropy boosting. Chaos: An Interdisciplinary Journal of Nonlinear Science, 33(8), 2023.
  115. Jared Elinger. Information Theoretic Causality Measures For Parameter Estimation and System Identification. PhD thesis, Georgia Institute of Technology, 2020.
  116. Thomas M Cover. Elements of information theory. John Wiley & Sons, 1999.
  117. Richard Ernest Bellman. Dynamic programming treatment of the traveling salesman problem. RAND Corporation, 1961.
  118. Measuring the potential utility of seasonal climate predictions. Geophysical research letters, 31(22), 2004.
  119. Richard Kleeman. Information theory and dynamical system predictability. Entropy, 13(3):612–649, 2011.
  120. Quantifying uncertainty for predictions with model error in non-Gaussian systems with intermittency. Nonlinearity, 25(9):2543, 2012.
  121. Non-gaussian test models for prediction and state estimation with model errors. Chinese Annals of Mathematics, Series B, 34(1):29–64, 2013.
  122. Xue Ying. An overview of overfitting and its solutions. In Journal of physics: Conference series, volume 1168, page 022022. IOP Publishing, 2019.
  123. Discovering governing equations from data by sparse identification of nonlinear dynamical systems. Proceedings of the national academy of sciences, 113(15):3932–3937, 2016.
  124. Nan Chen. Learning nonlinear turbulent dynamics from partial observations via analytically solvable conditional statistics. Journal of Computational Physics, 418:109635, 2020.
  125. Physics constrained nonlinear regression models for time series. Nonlinearity, 26(1):201, 2012.
  126. An ensemble Kalman filter for statistical estimation of physics constrained nonlinear regression models. Journal of Computational Physics, 257:782–812, 2014.
  127. An ensemble Kalman-Bucy filter for continuous data assimilation. Meteorologische Zeitschrift, 21(3):213, 2012.
  128. Edward N Lorenz. Formulation of a low-order model of a moist general circulation. Journal of Atmospheric Sciences, 41(12):1933–1945, 1984.
  129. Edward N Lorenz. Irregularity: A fundamental property of the atmosphere. Tellus A, 36(2):98–110, 1984.
  130. Stochastic parameterizing manifolds and non-Markovian reduced equations: stochastic manifolds for nonlinear SPDEs II. Springer, 2014.
  131. Post-processing finite-horizon parameterizing manifolds for optimal control of nonlinear parabolic pdes. In 2016 IEEE 55th Conference on Decision and Control (CDC), pages 1411–1416. IEEE, 2016.
  132. Concrete ensemble Kalman filters with rigorous catastrophic filter divergence. Proceedings of the National Academy of Sciences, 112(34):10589–10594, 2015.
  133. Data assimilation: methods, algorithms, and applications. SIAM, 2016.
  134. Edward N Lorenz. Predictability: A problem partly solved. In Proc. Seminar on predictability, volume 1. Reading, 1996.
  135. Exploiting local low dimensionality of the atmospheric dynamics for efficient ensemble kalman filtering. arXiv preprint physics/0203058, 3, 2002.
  136. Daniel S Wilks. Effects of stochastic parametrizations in the Lorenz’96 system. Quarterly Journal of the Royal Meteorological Society: A journal of the atmospheric sciences, applied meteorology and physical oceanography, 131(606):389–407, 2005.
  137. Stochastic parametrizations and model uncertainty in the lorenz’96 system. Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences, 371(1991):20110479, 2013.
  138. Recurrent neural networks. Design and Applications, 5(64-67):2, 2001.
  139. Long short-term memory. Neural computation, 9(8):1735–1780, 1997.
  140. Attention is all you need. Advances in neural information processing systems, 30, 2017.
  141. On the opportunities and risks of foundation models. arXiv preprint arXiv:2108.07258, 2021.
Citations (3)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets