Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
129 tokens/sec
GPT-4o
28 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

DySLIM: Dynamics Stable Learning by Invariant Measure for Chaotic Systems (2402.04467v2)

Published 6 Feb 2024 in cs.LG and math.DS

Abstract: Learning dynamics from dissipative chaotic systems is notoriously difficult due to their inherent instability, as formalized by their positive Lyapunov exponents, which exponentially amplify errors in the learned dynamics. However, many of these systems exhibit ergodicity and an attractor: a compact and highly complex manifold, to which trajectories converge in finite-time, that supports an invariant measure, i.e., a probability distribution that is invariant under the action of the dynamics, which dictates the long-term statistical behavior of the system. In this work, we leverage this structure to propose a new framework that targets learning the invariant measure as well as the dynamics, in contrast with typical methods that only target the misfit between trajectories, which often leads to divergence as the trajectories' length increases. We use our framework to propose a tractable and sample efficient objective that can be used with any existing learning objectives. Our Dynamics Stable Learning by Invariant Measure (DySLIM) objective enables model training that achieves better point-wise tracking and long-term statistical accuracy relative to other learning objectives. By targeting the distribution with a scalable regularization term, we hope that this approach can be extended to more complex systems exhibiting slowly-variant distributions, such as weather and climate models.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (135)
  1. TensorFlow: Large-scale machine learning on heterogeneous systems, 2015. URL https://www.tensorflow.org/. Software available from tensorflow.org.
  2. Operator-theoretic framework for forecasting nonlinear time series with kernel analog techniques. Physica D: Nonlinear Phenomena, 409:132520, 2020.
  3. Nonlinear model order reduction based on local reduced-order bases. International Journal for Numerical Methods in Engineering, 92(10):891–916, 2012.
  4. 2022 review of data-driven plasma science. arXiv preprint arXiv:2205.15832, 2022.
  5. Missing point estimation in models described by proper orthogonal decomposition. IEEE Transactions on Automatic Control, 53(10):2237–2251, 2008. doi:10.1109/TAC.2008.2006102.
  6. The dynamics of coherent structures in the wall region of a turbulent boundary layer. Journal of Fluid Mechanics, 192:115–173, 1988. doi:10.1017/S0022112088001818.
  7. Learning dynamical systems from partial observations, 2019.
  8. Learning data-driven discretizations for partial differential equations. Proceedings of the National Academy of Sciences, 116(31):15344–15349, July 2019. ISSN 0027-8424, 1091-6490. doi:10.1073/pnas.1814058116. URL https://pnas.org/doi/full/10.1073/pnas.1814058116.
  9. On a new multivariate two-sample test. Journal of multivariate analysis, 88(1):190–206, 2004.
  10. An ‘empirical interpolation’ method: application to efficient reduced-basis discretization of partial differential equations. Comptes Rendus Mathematique, 339(9):667–672, 2004. ISSN 1631-073X. doi:https://doi.org/10.1016/j.crma.2004.08.006. URL https://www.sciencedirect.com/science/article/pii/S1631073X04004248.
  11. Accurate medium-range global weather forecasting with 3d neural networks. Nature, 619(7970):533–538, 2023.
  12. Demystifying MMD GANs. In International Conference on Learning Representations, 2018. URL https://openreview.net/forum?id=r1lUOzWCW.
  13. Least-squares shadowing sensitivity analysis of chaotic flow around a two-dimensional airfoil. AIAA Journal, 56(2):658–672, 2018.
  14. Bollt, E. On explaining the surprising success of reservoir computing forecaster of chaos? the universal machine learning dynamical system with contrast to var and dmd. Chaos: An Interdisciplinary Journal of Nonlinear Science, 31(1), 2021.
  15. Spherical Fourier neural operators: Learning stable dynamics on the sphere. arXiv preprint arXiv:2306.03838, 2023.
  16. Neural ideal large eddy simulation: Modeling turbulence with neural stochastic differential equations. In Thirty-seventh Conference on Neural Information Processing Systems, 2023. URL https://openreview.net/forum?id=x6cOcxRnxG.
  17. Learning dynamics on invariant measures using PDE-constrained optimization. Chaos: An Interdisciplinary Journal of Nonlinear Science, 33(6), 2023.
  18. JAX: composable transformations of Python+NumPy programs, 2018. URL http://github.com/google/jax.
  19. Message passing neural PDE solvers. arXiv preprint arXiv:2202.03376, 2022.
  20. Brezis, H. Remarks on the Monge–Kantorovich problem in the discrete setting. Comptes Rendus. Mathématique, 356(2):207–213, 2018.
  21. FC-based shock-dynamics solver with neural-network localized artificial-viscosity assignment. arXiv:2111.01315 [cs, math], November 2021. arXiv: 2111.01315.
  22. Data-driven science and engineering: Machine learning, dynamical systems, and control. Cambridge University Press, 2022.
  23. Spectral Methods: Evolution to Complex Geometries and Applications to Fluid Dynamics (Scientific Computation). Springer-Verlag, Berlin, Heidelberg, 2007. ISBN 3540307273.
  24. Nonlinear model reduction via discrete empirical interpolation. SIAM Journal on Scientific Computing, 32(5):2737–2764, 2010. doi:10.1137/090766498. URL https://doi.org/10.1137/090766498.
  25. Training deep nets with sublinear memory cost. arXiv preprint arXiv:1604.06174, 2016.
  26. DeePKS: A comprehensive data-driven approach toward chemically accurate density functional theory. Journal of Chemical Theory and Computation, 17(1):170–181, 2020.
  27. A short review on model order reduction based on proper generalized decomposition. Archives of Computational Methods in Engineering, 18(4):395–404, 2011.
  28. An algorithm for the machine calculation of complex Fourier series. Math. Comput., 19(90):297–301, 1965. ISSN 00255718, 10886842. URL http://www.jstor.org/stable/2003354.
  29. Cuturi, M. Sinkhorn distances: Lightspeed computation of optimal transport. Advances in neural information processing systems, 26, 2013.
  30. Optimal transport tools (ott): A JAX toolbox for all things Wasserstein. arXiv preprint arXiv:2201.12324, 2022.
  31. Learning to correct spectral methods for simulating turbulent flows. arXiv preprint arXiv:2207.00556, 2022.
  32. Training generative neural networks via maximum mean discrepancy optimization. arXiv preprint arXiv:1505.03906, 2015.
  33. Long-term prediction of chaotic systems with machine learning. Physical Review Research, 2(1):012080, 2020.
  34. A multiscale neural network based on hierarchical nested bases. Research in the Mathematical Sciences, 6, Mar. 2019. ISSN 2197-9847. doi:10.1007/s40687-019-0183-3.
  35. Predicting chaotic time series. Phys. Rev. Lett., 59:845–848, Aug 1987. doi:10.1103/PhysRevLett.59.845. URL https://link.aps.org/doi/10.1103/PhysRevLett.59.845.
  36. Ferrario, B. Invariant measures for a stochastic kuramoto–sivashinsky equation. Stochastic analysis and applications, 26(2):379–407, 2008.
  37. Interpolating between optimal transport and MMD using sinkhorn divergences. In The 22nd International Conference on Artificial Intelligence and Statistics, pp.  2681–2690. PMLR, 2019.
  38. A posteriori learning for quasi-geostrophic turbulence parametrization. arXiv, April 2022.
  39. Galerkin, B. G. Series occurring in various questions concerning the elastic equilibrium of rods and plates. Vestnik Inzhenernov i Tekhnikov, 19:897–908, 1915.
  40. Operator inference for non-intrusive model reduction with nonlinear manifolds. arXiv:2205.02304 [math.NA], May 2022. URL https://arxiv.org/abs/2205.02304. arXiv: 2205.02304.
  41. Learning generative models with sinkhorn divergences. In International Conference on Artificial Intelligence and Statistics, pp.  1608–1617. PMLR, 2018.
  42. Data-driven prediction in dynamical systems: recent developments. Philosophical Transactions of the Royal Society A, 380(2229):20210213, 2022.
  43. Strictly proper scoring rules, prediction, and estimation. Journal of the American statistical Association, pp. 359–378, 2007.
  44. Generative adversarial networks. Communications of the ACM, 63(11):139–144, 2020.
  45. A kernel method for the two-sample-problem. Advances in neural information processing systems, 19, 2006.
  46. A kernel two-sample test. The Journal of Machine Learning Research, 13(1):723–773, 2012.
  47. Learning dynamics by reservoir computing (in memory of prof. pavol brunovskỳ). Journal of Dynamics and Differential Equations, pp.  1–26, 2022.
  48. Array programming with NumPy. Nature, 585(7825):357–362, September 2020. doi:10.1038/s41586-020-2649-2. URL https://doi.org/10.1038/s41586-020-2649-2.
  49. Hawkins, J. Attractors in dynamical systems. Ergodic dynamics: From basic theory to applications, pp. 27–39, 2021.
  50. Flax: A neural network library and ecosystem for JAX, 2023. URL http://github.com/google/flax.
  51. xarray: N-D labeled arrays and datasets in Python. Journal of Open Research Software, 5(1), 2017. doi:10.5334/jors.148. URL https://doi.org/10.5334/jors.148.
  52. Hunter, J. D. Matplotlib: A 2d graphics environment. Computing in Science & Engineering, 9(3):90–95, 2007. doi:10.1109/MCSE.2007.55.
  53. Pushing the limit of molecular dynamics with ab initio accuracy to 100 million atoms with machine learning. In SC20: International conference for high performance computing, networking, storage and analysis, pp.  1–14. IEEE, 2020.
  54. Training neural operators to preserve invariant measures of chaotic attractors. arXiv preprint arXiv:2306.01187, 2023.
  55. Data-driven discovery of Koopman eigenfunctions for control. Machine Learning: Science and Technology, 2(3):035023, 2021.
  56. Kantorovich, L. V. On the translocation of masses. In Dokl. Akad. Nauk. USSR (NS), volume 37, pp.  199–201, 1942.
  57. Keisler, R. Forecasting global weather with graph neural networks. arXiv preprint arXiv:2202.07575, 2022.
  58. Machine learning–accelerated computational fluid dynamics. Proc. Natl. Acad. Sci. U. S. A., 118(21), May 2021. URL https://www.pnas.org/content/118/21/e2101784118.
  59. Neural general circulation models. arXiv preprint arXiv:2311.07222, 2023.
  60. Koopman, B. O. Hamiltonian systems and transformation in hilbert space. Proceedings of the National Academy of Sciences, 17(5):315–318, 1931.
  61. Kuramoto, Y. Diffusion-induced chaos in reaction systems. Progress of Theoretical Physics Supplement, 64:346–367, 1978.
  62. Kutta, W. Beitrag zur näherungsweisen Integration totaler Differentialgleichungen. Teubner, 1901.
  63. Graphcast: Learning skillful medium-range global weather forecasting. arXiv preprint arXiv:2212.12794, 2022.
  64. Professor forcing: A new algorithm for training recurrent networks. Advances in neural information processing systems, 29, 2016.
  65. MMD GAN: Towards deeper understanding of moment matching network. Advances in neural information processing systems, 30, 2017.
  66. Generative moment matching networks. In International conference on machine learning, pp. 1718–1727. PMLR, 2015.
  67. Multipole graph neural operator for parametric partial differential equations. In Proceedings of the 34th International Conference on Neural Information Processing Systems, NIPS’20, Red Hook, NY, USA, 2020. Curran Associates Inc. ISBN 9781713829546.
  68. Fourier Neural Operator for Parametric Partial Differential Equations. arXiv:2010.08895 [cs, math], May 2021. URL http://arxiv.org/abs/2010.08895. arXiv: 2010.08895.
  69. Learning chaotic dynamics in dissipative systems. Advances in Neural Information Processing Systems, 35:16768–16781, 2022.
  70. Learned Turbulence Modelling with Differentiable Fluid Solvers. arXiv:2202.06988 [physics], February 2022. URL http://arxiv.org/abs/2202.06988. arXiv: 2202.06988.
  71. Learning deep kernels for non-parametric two-sample tests. In Proceedings of the 37th International Conference on Machine Learning, volume 119, pp.  6316–6326. PMLR, 2020.
  72. Lorenz, E. N. Deterministic nonperiodic flow. Journal of atmospheric sciences, 20(2):130–141, 1963.
  73. Data-based stochastic model reduction for the Kuramoto–Sivashinsky equation. Physica D: Nonlinear Phenomena, 340:46–57, 2017.
  74. Learning nonlinear operators via DeepONet based on the universal approximation theorem of operators. Nature Machine Intelligence, 3(3):218–229, 2021.
  75. The lorenz attractor is mixing. Communications in Mathematical Physics, 260:393–401, 2005.
  76. Uncovering turbulent plasma dynamics via deep learning from partial observations. Physical Review E, 104(2):025205, 2021.
  77. Nonlinear dynamics: A primer. Cambridge University Press, 2001.
  78. Scaling deep learning for materials discovery. Nature, pp.  1–6, 2023.
  79. On the difficulty of learning chaotic dynamics with RNNs. Advances in Neural Information Processing Systems, 35:11297–11312, 2022.
  80. Mishra, S. A machine learning framework for data driven acceleration of computations of differential equations. Mathematics in Engineering, 1(1):118–146, 2018. ISSN 2640-3501. doi:10.3934/Mine.2018.1.118. URL https://www.aimspress.com/article/doi/10.3934/Mine.2018.1.118.
  81. Müller, A. Integral probability metrics and their generating classes of functions. Advances in applied probability, 29(2):429–443, 1997.
  82. Physics-informed machine learning for modeling and control of dynamical systems. arXiv preprint arXiv:2306.13867, 2023.
  83. Obukhov, A. Kolmogorov flow and laboratory simulation of it. Russ. Math. Surv, 38(4):113–126, 1983.
  84. pandas development team, T. pandas-dev/pandas: Pandas, February 2020. URL https://doi.org/10.5281/zenodo.3509134.
  85. The route to chaos for the kuramoto-sivashinsky equation. Theoretical and Computational Fluid Dynamics, 3(1):15–42, 1991.
  86. Normalizing flows for probabilistic modeling and inference. The Journal of Machine Learning Research, 22(1):2617–2680, 2021.
  87. On the difficulty of training recurrent neural networks. In International conference on machine learning, pp. 1310–1318. Pmlr, 2013.
  88. Using machine learning to replicate chaotic attractors and calculate Lyapunov exponents from data. Chaos: An Interdisciplinary Journal of Nonlinear Science, 27(12), 2017.
  89. Fourcastnet: A global data-driven high-resolution weather model using adaptive fourier neural operators. arXiv preprint arXiv:2202.11214, 2022.
  90. Pearson, K. Liii. on lines and planes of closest fit to systems of points in space. The London, Edinburgh, and Dublin Philosophical Magazine and Journal of Science, 2(11):559–572, 1901. doi:10.1080/14786440109462720. URL https://doi.org/10.1080/14786440109462720.
  91. Peherstorfer, B. Breaking the kolmogorov barrier with nonlinear model reduction. Notices of the American Mathematical Society, 69(5):725–733, 2022.
  92. Computational optimal transport: With applications to data science. Foundations and Trends® in Machine Learning, 11(5-6):355–607, 2019.
  93. Modeling of dynamical systems through deep learning. Biophysical Reviews, 12(6):1311–1320, 2020.
  94. Searching for activation functions. In 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Workshop Track Proceedings. OpenReview.net, 2018. URL https://openreview.net/forum?id=Hkuq2EkPf.
  95. On wasserstein two-sample testing and related families of nonparametric tests. Entropy, 19(2):47, 2017.
  96. Gaussian processes for machine learning, volume 1. Springer, 2006.
  97. Variational inference with normalizing flows. In International conference on machine learning, pp. 1530–1538. PMLR, 2015.
  98. Machine learning in nonlinear dynamical systems. Resonance, 26(7):953–970, 2021.
  99. Runge, C. Über die numerische auflösung von differentialgleichungen. Mathematische Annalen, 46(2):167–178, 1895.
  100. Learning to simulate complex physics with graph networks. In International conference on machine learning, pp. 8459–8468. PMLR, 2020.
  101. Santambrogio, F. Optimal transport for applied mathematicians. Birkäuser, NY, 55(58-63):94, 2015.
  102. Schmid, P. J. Dynamic mode decomposition of numerical and experimental data. Journal of Fluid Mechanics, 656:5–28, 2010. doi:10.1017/S0022112010001217.
  103. Schmid, P. J. Dynamic mode decomposition and its variants. Annual Review of Fluid Mechanics, 54:225–254, 2022.
  104. MMD aggregated two-sample test. Journal of Machine Learning Research, 24(194):1–81, 2023.
  105. Equivalence of distance-based and RKHS-based statistics in hypothesis testing. The Annals of Statistics, 41(5):2263 – 2291, 2013. doi:10.1214/13-AOS1140. URL https://doi.org/10.1214/13-AOS1140.
  106. Operator learning with neural fields: Tackling PDEs on general geometries. arXiv preprint arXiv:2306.07266, 2023.
  107. Autoregressive quantile flows for predictive uncertainty estimation. arXiv preprint arXiv:2112.04643, 2021.
  108. Semi-autoregressive energy flows: exploring likelihood-free training of normalizing flows. In International Conference on Machine Learning, pp. 31732–31753. PMLR, 2023.
  109. Metrizing weak convergence with maximum mean discrepancies. Journal of Machine Learning Research, 24(184):1–20, 2023.
  110. Sivashinsky, G. I. Nonlinear analysis of hydrodynamic instability in laminar flames—i. derivation of basic equations. In Dynamics of Curved Fronts, pp.  459–488. Elsevier, 1988.
  111. On integral probability metrics,\\\backslash\phi-divergences and binary classification. arXiv preprint arXiv:0901.2698, 2009.
  112. Hilbert space embeddings and metrics on probability measures. The Journal of Machine Learning Research, 11:1517–1561, 2010.
  113. Learned Coarse Models for Efficient Turbulence Simulation. arXiv:2112.15275 [physics], January 2022. URL http://arxiv.org/abs/2112.15275. arXiv: 2112.15275.
  114. Strogatz, S. H. Nonlinear dynamics and chaos with student solutions manual: With applications to physics, biology, chemistry, and engineering. CRC press, 2018.
  115. Dynamical systems and numerical analysis, volume 2. Cambridge University Press, 1998.
  116. Stuart, A. M. Numerical analysis of dynamical systems. Acta numerica, 3:467–572, 1994.
  117. Testing for equal distributions in high dimension. InterStat, 5(16.10):1249–1272, 2004.
  118. Temam, R. Infinite-dimensional dynamical systems in mechanics and physics, volume 68. Springer Science & Business Media, 2012.
  119. Minimax estimation of maximum mean discrepancy with radial kernels. In Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., and Garnett, R. (eds.), Advances in Neural Information Processing Systems, volume 29, 2016. URL https://proceedings.neurips.cc/paper_files/paper/2016/file/5055cbf43fac3f7e2336b27310f0b9ef-Paper.pdf.
  120. Factorized Fourier Neural Operators. arXiv:2111.13802 [cs], November 2021. arXiv: 2111.13802.
  121. Trefethen, L. N. Spectral Methods in MATLAB. Society for industrial and applied mathematics (SIAM), 2000.
  122. Tucker, W. The Lorenz attractor exists. Comptes Rendus de l’Académie des Sciences-Series I-Mathematics, 328(12):1197–1202, 1999.
  123. Tucker, W. A rigorous ODE solver and Smale’s 14th problem. Foundations of Computational Mathematics, 2:53–117, 2002.
  124. Solver-in-the-loop: Learning from differentiable physics to interact with iterative pde-solvers. In Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., and Lin, H. (eds.), Advances in Neural Information Processing Systems, volume 33, pp.  6111–6122. Curran Associates, Inc., 2020. URL https://proceedings.neurips.cc/paper_files/paper/2020/file/43e4e6a6f341e00671e123714de019a8-Paper.pdf.
  125. SciPy 1.0: Fundamental Algorithms for Scientific Computing in Python. Nature Methods, 17:261–272, 2020. doi:10.1038/s41592-019-0686-2.
  126. Data-driven forecasting of high-dimensional chaotic systems with long short-term memory networks. Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences, 474(2213):20170844, 2018.
  127. Backpropagation algorithms and reservoir computing in recurrent neural networks for the forecasting of complex spatiotemporal dynamics. Neural Networks, 126:191–217, 2020.
  128. Debias coarsely, sample conditionally: Statistical downscaling through optimal transport and probabilistic diffusion models. In Thirty-seventh Conference on Neural Information Processing Systems, 2023a. URL https://openreview.net/forum?id=5NxJuc0T1P.
  129. Evolve smoothly, fit consistently: Learning smooth latent dynamics for advection-dominated systems. In The Eleventh International Conference on Learning Representations, 2023b.
  130. Least squares shadowing sensitivity analysis of chaotic limit cycle oscillations. Journal of Computational Physics, 267:210–224, 2014.
  131. Waskom, M. L. seaborn: statistical data visualization. Journal of Open Source Software, 6(60):3021, 2021. doi:10.21105/joss.03021. URL https://doi.org/10.21105/joss.03021.
  132. ACE: A fast, skillful learned global atmospheric model for climate prediction. arXiv preprint arXiv:2310.02074, 2023.
  133. Gibbsian dynamics and invariant measures for stochastic dissipative PDEs. Journal of Statistical Physics, 108:1125–1156, 2002.
  134. Willcox, K. Unsteady flow sensing and estimation via the gappy proper orthogonal decomposition. Computers & Fluids, 35(2):208–226, 2006. ISSN 0045-7930. doi:https://doi.org/10.1016/j.compfluid.2004.11.006. URL https://www.sciencedirect.com/science/article/pii/S0045793005000113.
  135. Deep density: circumventing the Kohn-Sham equations via symmetry preserving neural networks. Journal of Computational Physics, 443:110523, 2021.
Citations (6)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com