Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash 88 tok/s
Gemini 2.5 Pro 40 tok/s Pro
GPT-5 Medium 34 tok/s
GPT-5 High 36 tok/s Pro
GPT-4o 84 tok/s
GPT OSS 120B 463 tok/s Pro
Kimi K2 191 tok/s Pro
2000 character limit reached

Learning About Structural Errors in Models of Complex Dynamical Systems (2401.00035v2)

Published 29 Dec 2023 in physics.comp-ph, cs.LG, and math.DS

Abstract: Complex dynamical systems are notoriously difficult to model because some degrees of freedom (e.g., small scales) may be computationally unresolvable or are incompletely understood, yet they are dynamically important. For example, the small scales of cloud dynamics and droplet formation are crucial for controlling climate, yet are unresolvable in global climate models. Semi-empirical closure models for the effects of unresolved degrees of freedom often exist and encode important domain-specific knowledge. Building on such closure models and correcting them through learning the structural errors can be an effective way of fusing data with domain knowledge. Here we describe a general approach, principles, and algorithms for learning about structural errors. Key to our approach is to include structural error models inside the models of complex systems, for example, in closure models for unresolved scales. The structural errors then map, usually nonlinearly, to observable data. As a result, however, mismatches between model output and data are only indirectly informative about structural errors, due to a lack of labeled pairs of inputs and outputs of structural error models. Additionally, derivatives of the model may not exist or be readily available. We discuss how structural error models can be learned from indirect data with derivative-free Kalman inversion algorithms and variants, how sparsity constraints enforce a "do no harm" principle, and various ways of modeling structural errors. We also discuss the merits of using non-local and/or stochastic error models. In addition, we demonstrate how data assimilation techniques can assist the learning about structural errors in non-ergodic systems. The concepts and algorithms are illustrated in two numerical examples based on the Lorenz-96 system and a human glucose-insulin model.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (115)
  1. Climate goals and computing the future of clouds, Nature Climate Change 7 (2017) 3–5.
  2. J. Smagorinsky, General circulation experiments with the primitive equations. I. The basic experiment, Mon. Wea. Rev. 91 (1963) 99–164.
  3. A. Holtslag, C.-H. Moeng, Eddy diffusivity and countergradient transport in the convective atmospheric boundary layer, Journal of the Atmospheric Sciences 48 (1991) 1690–1698.
  4. Reynolds averaged turbulence modelling using deep neural networks with embedded invariance, Journal of Fluid Mechanics 807 (2016) 155–166.
  5. J. Weatheritt, R. Sandberg, A novel evolutionary algorithm applied to algebraic modifications of the RANS stress–strain relationship, Journal of Computational Physics 325 (2016) 22–37.
  6. Deep learning to represent subgrid processes in climate models, Proc. Natl. Acad. Sci. (2018).
  7. Subgrid modelling for two-dimensional turbulence using neural networks, Journal of Fluid Mechanics 858 (2019) 122–144.
  8. Turbulence modeling in the age of data, Annual Review of Fluid Mechanics 51 (2019) 357–377.
  9. RANS turbulence model development using CFD-driven machine learning, Journal of Computational Physics 411 (2020) 109413.
  10. Discovery of algebraic Reynolds-stress models using sparse symbolic regression, Flow, Turbulence and Combustion 104 (2020) 579–603.
  11. Machine learning for fluid mechanics, Ann. Rev. Fluid Mech. 52 (2020) 477–508.
  12. M. C. Kennedy, A. O’Hagan, Bayesian calibration of computer models, Journal of the Royal Statistical Society: Series B (Statistical Methodology) 63 (2001) 425–464.
  13. Combining field data and computer simulations for calibration and prediction, SIAM Journal on Scientific Computing 26 (2004) 448–466.
  14. Bayesian validation of computer models, Technometrics 51 (2009) 439–451.
  15. J. Brynjarsdòttir, A. O’Hagan, Learning about physical parameters: The importance of model discrepancy, Inverse problems 30 (2014) 114007.
  16. Error modeling for surrogates of dynamical systems using machine learning, International Journal for Numerical Methods in Engineering 112 (2017) 1801–1827.
  17. Iterative updating of model error for Bayesian inversion, Inverse Problems 34 (2018) 025008.
  18. M. Strong, J. E. Oakley, When is a model good enough? Deriving the expected value of model improvement via specifying internal model discrepancies, SIAM/ASA Journal on Uncertainty Quantification 2 (2014) 106–125.
  19. Y. He, D. Xiu, Numerical strategy for model correction using physical constraints, Journal of Computational Physics 313 (2016) 617–634.
  20. Embedded model error representation for Bayesian model calibration, International Journal for Uncertainty Quantification 9 (2019).
  21. Modeling structural uncertainties in Reynolds-averaged computations of shock/boundary layer interactions, in: 49th AIAA Aerospace sciences meeting including the new horizons forum and aerospace exposition, 2011, p. 479.
  22. T. A. Oliver, R. D. Moser, Bayesian uncertainty quantification applied to RANS turbulence models, in: Journal of Physics: Conference Series, volume 318, IOP Publishing, 2011, p. 042032.
  23. Bayesian uncertainty analysis with applications to turbulence modeling, Reliability Engineering & System Safety 96 (2011) 1137–1149.
  24. Quantifying and reducing model-form uncertainties in Reynolds-averaged Navier–Stokes simulations: A data-driven, physics-informed Bayesian approach, Journal of Computational Physics 324 (2016) 115–136.
  25. P. Pernot, F. Cailliez, A critical review of statistical calibration/prediction models handling data inconsistency and model inadequacy, AIChE Journal 63 (2017) 4642–4665.
  26. Physics-informed machine learning approach for reconstructing Reynolds stress modeling discrepancies based on DNS data, Physical Review Fluids 2 (2017) 034603.
  27. Physics-informed machine learning approach for augmenting turbulence models: A comprehensive framework, Physical Review Fluids 3 (2018) 074602.
  28. Sparse dynamics for partial differential equations, Proceedings of the National Academy of Sciences 110 (2013) 6634–6639.
  29. Data-driven discovery of partial differential equations, Science Advances 3 (2017) e1602614.
  30. H. Schaeffer, Learning partial differential equations via data discovery and sparse optimization, Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences 473 (2017) 20160446.
  31. L. Zanna, T. Bolton, Data-driven equation discovery of ocean mesoscale closures, Geophysical Research Letters 47 (2020) e2020GL088376.
  32. Discovery of physics from data: Universal laws and discrepancies, Front. Artif. Intell. 3 (2020) 25.
  33. Rediscovering orbital mechanics with machine learning (2022).
  34. N. D. Brenowitz, C. S. Bretherton, Prognostic validation of a neural network unified physics parameterization, Geophysical Research Letters 45 (2018) 6289–6298.
  35. L. Zanna, T. Bolton, Deep learning of unresolved turbulent ocean processes in climate models, Deep Learning for the Earth Sciences: A Comprehensive Approach to Remote Sensing, Climate Science, and Geosciences (2021) 298–306.
  36. Correcting coarse-grid weather and climate models by machine learning from global storm-resolving simulations, J. Adv. Model. Earth Sys. 14 (2022) e2021MS002794.
  37. Mathematical and physical ideas for climate science, Rev. Geophys. 52 (2014) 809–859.
  38. Stochastic climate theory and modeling, Wiley Interdisciplinary Reviews: Climate Change 6 (2015) 63–78.
  39. T. Palmer, Stochastic weather and climate models, Nature Reviews Physics 1 (2019) 463–471.
  40. Learning nonlocal constitutive models with neural networks, Computer Methods in Applied Mechanics and Engineering 384 (2021) 113927.
  41. Learned discretizations for passive scalar advection in a two-dimensional turbulent flow, Phys. Rev. Fluids 6 (2021) 064605.
  42. Analysis and approximation of nonlocal diffusion problems with volume constraints, SIAM Review 54 (2012) 667–696.
  43. M. D’Elia, M. Gunzburger, Identification of the diffusion parameter in nonlocal steady diffusion problems, Applied Mathematics & Optimization 73 (2016) 227–249.
  44. nPINNs: nonlocal physics-informed neural networks for a parametrized nonlocal universal Laplacian operator. algorithms and applications, Journal of Computational Physics 422 (2020) 109760.
  45. Model reduction with memory and the machine learning of dynamical systems, arXiv preprint arXiv:1808.04258 (2018).
  46. Recurrent neural network closure of parametric POD-Galerkin reduced-order models based on the Mori-Zwanzig formalism, Journal of Computational Physics 410 (2020) 109402.
  47. K. K. Lin, F. Lu, Data-driven model reduction, Wiener projections, and the Koopman-Mori-Zwanzig formalism, Journal of Computational Physics 424 (2021) 109864.
  48. A.-T. G. Charalampopoulos, T. P. Sapsis, Machine-learning energy-preserving nonlocal closures for turbulent fluid flows and inertial tracers, arXiv preprint arXiv:2102.07639 (2021).
  49. Discovering governing equations from data by sparse identification of nonlinear dynamical systems, Proceedings of the National Academy of Sciences 113 (2016) 3932–3937.
  50. M. Levine, A. Stuart, A framework for machine learning of model error in dynamical systems, Communications of the American Mathematical Society 2 (2022) 283–344.
  51. Learning stochastic closures using ensemble kalman inversion, Transactions of Mathematics and Its Applications 5 (2021) tnab003.
  52. Ensemble kalman method for learning turbulence models from indirect observation data, Journal of Fluid Mechanics 949 (2022) A26.
  53. E. N. Lorenz, Predictability: A problem partly solved, in: Proc. Seminar on predictability, volume 1, 1996.
  54. I. Fatkullin, E. Vanden-Eijnden, A computational strategy for multiscale systems with applications to lorenz 96 model, Journal of Computational Physics 200 (2004) 605–638.
  55. Data assimilation as a learning tool to infer ordinary differential equation representations of dynamical models, Nonlinear Processes in Geophysics 26 (2019) 143–162.
  56. Autodifferentiable ensemble Kalman filters, SIAM Journal on Mathematics of Data Science 4 (2022) 801–833.
  57. Deep neural networks for data-driven LES closure models, Journal of Computational Physics 398 (2019) 108910.
  58. Interpreting and stabilizing machine-learning parametrizations of convection, Journal of the Atmospheric Sciences 77 (2020) 4357–4375.
  59. Use of neural networks for stable, accurate and physically consistent parameterization of subgrid atmospheric processes with good performance at reduced precision, Geophysical Research Letters 48 (2021) e2020GL091363.
  60. Enforcing analytic constraints in neural networks emulating physical systems, Physical Review Letters 126 (2021) 098302.
  61. Neural lander: Stable drone landing control using learned dynamics, in: 2019 International Conference on Robotics and Automation (ICRA), IEEE, 2019, pp. 9784–9790.
  62. Bayesian inference for a discretely observed stochastic kinetic model, Statistics and Computing 18 (2008) 125–135.
  63. Earth system modeling 2.0: A blueprint for models that learn from observations and targeted high-resolution simulations, Geophysical Research Letters 44 (2017) 12–396.
  64. Parameter estimation for partially observed hypoelliptic diffusions, Journal of the Royal Statistical Society: Series B (Statistical Methodology) 71 (2009) 49–73.
  65. X.-L. Meng, D. Van Dyk, The EM algorithm—an old folk-song sung to a fast new tune, Journal of the Royal Statistical Society: Series B (Statistical Methodology) 59 (1997) 511–567.
  66. Ensemble inference methods for models with noisy and expensive likelihoods, SIAM Journal on Applied Dynamical Systems 21 (2022) 1539–1572.
  67. Ensemble Kalman inversion for sparse learning of dynamical systems from time-averaged data, Journal of Computational Physics 470 (2022) 111559.
  68. Calibrate, emulate, sample, Journal of Computational Physics 424 (2021) 109716.
  69. Model selection of chaotic systems from data with hidden variables using sparse data assimilation, Chaos: An Interdisciplinary Journal of Nonlinear Science 32 (2022).
  70. Combining data assimilation and machine learning to infer unresolved scale parametrization, Philosophical Transactions of the Royal Society A 379 (2021) 20200086.
  71. D. L. Donoho, Compressed sensing, IEEE Transactions on information theory 52 (2006) 1289–1306.
  72. B. Hamzi, H. Owhadi, Learning dynamical systems from data: a simple cross-validation perspective, part i: parametric kernel flows, Physica D: Nonlinear Phenomena 421 (2021) 132817.
  73. Learning dynamical systems from data: a simple cross-validation perspective, part ii: nonparametric kernel flows, preprint (2021).
  74. Learning dynamical systems from data: A simple cross-validation perspective, part iii: Irregularly-sampled time series, Physica D: Nonlinear Phenomena (2022) 133546.
  75. A. Rahimi, B. Recht, Random Features for Large-Scale Kernel Machines, in: J. Platt, D. Koller, Y. Singer, S. Roweis (Eds.), Advances in Neural Information Processing Systems, volume 20, Curran Associates, Inc., 2008.
  76. D. J. Wilkinson, Stochastic modelling for quantitative description of heterogeneous biological systems, Nature Reviews Genetics 10 (2009) 122–133.
  77. G. O. Roberts, O. Stramer, On inference for partially observed nonlinear diffusion models using the Metropolis–Hastings algorithm, Biometrika 88 (2001) 603–621.
  78. G. O. Roberts, O. Stramer, Langevin diffusions and Metropolis-Hastings algorithms, Methodology and computing in applied probability 4 (2002) 337–357.
  79. Data augmentation for diffusions, Journal of Computational and Graphical Statistics 22 (2013) 665–688.
  80. A tale of two time scales: Determining integrated volatility with noisy high-frequency data, Journal of the American Statistical Association 100 (2005) 1394–1411.
  81. G. A. Pavliotis, A. Stuart, Parameter estimation for multiscale diffusions, Journal of Statistical Physics 127 (2007) 741–781.
  82. Maximum likelihood drift estimation for multiscale diffusions, Stochastic Processes and their Applications 119 (2009) 3173–3210.
  83. Sparse learning of stochastic dynamical equations, The Journal of chemical physics 148 (2018) 241723.
  84. L. Zhang, Efficient estimation of stochastic volatility using noisy observations: A multi-scale approach, Bernoulli 12 (2006) 1019–1043.
  85. Nonparametric estimation of diffusions: a differential equations approach, Biometrika 99 (2012) 511–531.
  86. Parameter estimation for multiscale diffusions: An overview, Statistical Methods for Stochastic Differential Equations (2012) 429.
  87. Approximate Bayes learning of stochastic differential equations, Physical Review E 98 (2018) 022109.
  88. Drift estimation of multiscale diffusions based on filtered data, Foundations of Computational Mathematics (2021) 1–52.
  89. C. Honisch, R. Friedrich, Estimation of Kramers-Moyal coefficients at low sampling rates, Physical Review E 83 (2011) 066701.
  90. S. J. Lade, Finite sampling interval effects in Kramers–Moyal analysis, Physics Letters A 373 (2009) 3705–3709.
  91. Nonlinear stochastic modelling with Langevin regression, Proceedings of the Royal Society A 477 (2021) 20210092.
  92. A. Neumaier, T. Schneider, Estimation of parameters and eigenmodes of multivariate autoregressive models, ACM Trans. Math. Softw. 27 (2001) 27–57.
  93. T. Schneider, A. Neumaier, Algorithm 808: ARfit — A Matlab package for the estimation of parameters and eigenmodes of multivariate autoregressive models, ACM Trans. Math. Softw. 27 (2001) 58–65.
  94. Stochastic parametrizations and model uncertainty in the Lorenz’96 system, Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences 371 (2013) 20110479.
  95. F. Lu, Data-driven model reduction for stochastic Burgers equations, Entropy 22 (2020) 1360.
  96. N. Chen, A. J. Majda, Filtering nonlinear turbulent dynamical systems through conditional Gaussian statistics, Monthly Weather Review 144 (2016) 4885–4917.
  97. N. Chen, A. J. Majda, Conditional Gaussian systems for multiscale nonlinear stochastic systems: Prediction, state estimation and uncertainty quantification, Entropy 20 (2018) 509.
  98. Semiparametric drift and diffusion estimation for multiscale diffusions, Multiscale Modeling & Simulation 11 (2013) 442–473.
  99. Data-driven coarse graining in action: Modeling and prediction of complex systems, Physical Review E 92 (2015) 042139.
  100. A new framework for extracting coarse-grained models from time series with multiscale structure, Journal of Computational Physics 296 (2015) 314–328.
  101. K. Hasselmann, Stochastic climate models part i. theory, tellus 28 (1976) 473–485.
  102. C. Frankignoul, K. Hasselmann, Stochastic climate models, part ii application to sea-surface temperature anomalies and thermocline variability, Tellus 29 (1977) 289–305.
  103. C. Penland, T. Magorian, Prediction of Niño 3 sea surface temperatures using linear inverse modeling, Journal of Climate 6 (1993) 1067–1076.
  104. T. Schneider, S. M. Griffies, A conceptual framework for predictability studies, J. Climate 12 (1999) 3133–3155.
  105. K. Hasselmann, PIPs and POPs: The reduction of complex dynamical systems using principal interaction and oscillation patterns, Journal of Geophysical Research: Atmospheres 93 (1988) 11015–11021.
  106. Fourier neural operator for parametric partial differential equations, in: International Conference on Learning Representations, 2020.
  107. Learning nonlinear operators via deeponet based on the universal approximation theorem of operators, Nature Machine Intelligence 3 (2021) 218–229.
  108. N. H. Nelsen, A. M. Stuart, The random feature model for input-output maps between banach spaces, SIAM Journal on Scientific Computing 43 (2021) A3212–A3243.
  109. Neural operator: Learning maps between function spaces with applications to PDEs, Journal of Machine Learning Research 24 (2023) 1–97.
  110. Attention is all you need, Advances in neural information processing systems 30 (2017).
  111. Embedology, Journal of statistical Physics 65 (1991) 579–616.
  112. Computer model for mechanisms underlying ultradian oscillations of insulin and glucose, American Journal of Physiology-Endocrinology And Metabolism 260 (1991) E801–E809. Publisher: American Physiological Society Bethesda, MD.
  113. Ensemble kalman methods: A mean field perspective, arXiv preprint arXiv:2209.11371 (2022).
  114. Ensemble Kalman methods for inverse problems, Inverse Problems 29 (2013) 045001.
  115. Ensemble Kalman methods with constraints, Inverse Problems 35 (2019) 095007.
Citations (12)
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com