Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
143 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Equation Discovery with Bayesian Spike-and-Slab Priors and Efficient Kernels (2310.05387v2)

Published 9 Oct 2023 in cs.LG and stat.ML

Abstract: Discovering governing equations from data is important to many scientific and engineering applications. Despite promising successes, existing methods are still challenged by data sparsity and noise issues, both of which are ubiquitous in practice. Moreover, state-of-the-art methods lack uncertainty quantification and/or are costly in training. To overcome these limitations, we propose a novel equation discovery method based on Kernel learning and BAyesian Spike-and-Slab priors (KBASS). We use kernel regression to estimate the target function, which is flexible, expressive, and more robust to data sparsity and noises. We combine it with a Bayesian spike-and-slab prior -- an ideal Bayesian sparse distribution -- for effective operator selection and uncertainty quantification. We develop an expectation-propagation expectation-maximization (EP-EM) algorithm for efficient posterior inference and function estimation. To overcome the computational challenge of kernel regression, we place the function values on a mesh and induce a Kronecker product construction, and we use tensor algebra to enable efficient computation and optimization. We show the advantages of KBASS on a list of benchmark ODE and PDE discovery tasks.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (44)
  1. Latent force models. In Artificial Intelligence and Statistics, pages 9–16.
  2. Gaussian processes for Bayesian estimation in ordinary differential equations. In International Conference on Machine Learning, pages 1485–1493.
  3. Data-driven discovery of pdes in complex datasets. Journal of Computational Physics, 384:239–252.
  4. DeepMoD: Deep learning for model discovery in noisy data. Journal of Computational Physics, 428:109985.
  5. Discovering governing equations from data by sparse identification of nonlinear dynamical systems. Proceedings of the national academy of sciences, 113(15):3932–3937.
  6. Solving and learning nonlinear PDEs with Gaussian processes. arXiv preprint arXiv:2103.12959.
  7. Physics-informed learning of governing equations from scarce data. Nature communications, 12(1):1–13.
  8. Solving high frequency and multi-scale pdes with gaussian processes. arXiv preprint arXiv:2311.04465.
  9. Online Bayesian sparse learning with spike and slab priors. In 2020 IEEE International Conference on Data Mining (ICDM), pages 142–151. IEEE.
  10. Compiling machine learning programs via high-level tracing. Systems for Machine Learning, 4(9).
  11. Graepel, T. (2003). Solving noisy linear operator equations by Gaussian processes: Application to ordinary and partial differential equations. In ICML, pages 234–241.
  12. Learning unknown ODE models with Gaussian processes. In International Conference on Machine Learning, pages 1959–1968.
  13. Spike and slab variable selection: Frequentist and Bayesian strategies. The Annals of statistics, 33(2):730–773.
  14. Kolda, T. G. (2006). Multilinear operators for higher-order decompositions. Technical report, Sandia National Laboratories (SNL).
  15. Characterizing possible failure modes in physics-informed neural networks. Advances in Neural Information Processing Systems, 34:26548–26560.
  16. Learning partial differential equations for biological transport models from noisy spatio-temporal data. Proceedings of the Royal Society A, 476(2234):20190800.
  17. Scalable Gaussian process regression networks. In Proceedings of the Twenty-Ninth International Conference on International Joint Conferences on Artificial Intelligence, pages 2456–2462.
  18. Autoip: A united framework to integrate physics into gaussian processes. In International Conference on Machine Learning, pages 14210–14222. PMLR.
  19. Controversy in mechanistic modelling with Gaussian processes. Proceedings of Machine Learning Research, 37:1539–1547.
  20. MacKay, D. J. (2003). Information theory, inference and learning algorithms. Cambridge university press.
  21. A simple baseline for Bayesian uncertainty in deep learning. Advances in neural information processing systems, 32.
  22. Minka, T. P. (2000). Old and new matrix algebra useful for statistics.
  23. Minka, T. P. (2001a). Expectation propagation for approximate Bayesian inference. In Proceedings of the Seventeenth conference on Uncertainty in artificial intelligence, pages 362–369.
  24. Minka, T. P. (2001b). A family of algorithms for approximate Bayesian inference. PhD thesis, Massachusetts Institute of Technology.
  25. Bayesian variable selection in linear regression. Journal of the american statistical association, 83(404):1023–1032.
  26. Bayesian and l1 approaches for sparse unsupervised learning. In Proceedings of the 29th International Coference on International Conference on Machine Learning, pages 683–690.
  27. Lagrangian PINNs: A causality-conforming solution to failure modes of physics-informed neural networks. arXiv preprint arXiv:2205.02902.
  28. Machine learning of linear differential equations using Gaussian processes. Journal of Computational Physics, 348:683–693.
  29. Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. Journal of Computational Physics, 378:686–707.
  30. A scalable laplace approximation for neural networks. In 6th International Conference on Learning Representations, ICLR 2018-Conference Track Proceedings, volume 6. International Conference on Representation Learning.
  31. Data-driven discovery of partial differential equations. Science advances, 3(4):e1602614.
  32. Saatcci, Y. (2012). Scalable inference for structured Gaussian process models. PhD thesis, Citeseer.
  33. Schaeffer, H. (2017). Learning partial differential equations via data discovery and sparse optimization. Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences, 473(2197):20160446.
  34. Bayesian spline learning for equation discovery of nonlinear dynamics with quantified uncertainty. Advances in Neural Information Processing Systems, 35:6927–6940.
  35. Tipping, M. E. (2001). Sparse Bayesian learning and the relevance vector machine. Journal of machine learning research, 1(Jun):211–244.
  36. Graphical models, exponential families, and variational inference. Foundations and Trends® in Machine Learning, 1(1–2):1–305.
  37. Walker, A. M. (1969). On the asymptotic behaviour of posterior distributions. Journal of the Royal Statistical Society: Series B (Methodological), 31(1):80–88.
  38. Multi-fidelity high-order Gaussian processes for physical simulation. In International Conference on Artificial Intelligence and Statistics, pages 847–855. PMLR.
  39. ODIN: ODE-informed regression for parameter and state inference in time-continuous dynamical systems. In AAAI, pages 6364–6371.
  40. Fast Gaussian process based gradient matching for parameter identification in systems of nonlinear ODEs. In The 22nd International Conference on Artificial Intelligence and Statistics, pages 1351–1360. PMLR.
  41. Gaussian processes for machine learning, volume 2. MIT press Cambridge, MA.
  42. Kernel interpolation for scalable structured Gaussian processes (KISS-GP). In International conference on machine learning, pages 1775–1784. PMLR.
  43. Data-driven discovery of governing equations for fluid dynamics based on molecular simulation. Journal of Fluid Mechanics, 892.
  44. Scalable high-order Gaussian process regression. In The 22nd International Conference on Artificial Intelligence and Statistics, pages 2611–2620. PMLR.

Summary

We haven't generated a summary for this paper yet.

Github Logo Streamline Icon: https://streamlinehq.com

GitHub

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets