Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
143 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Bayesian Take on Gaussian Process Networks (2306.11380v4)

Published 20 Jun 2023 in stat.ML, cs.LG, and stat.ME

Abstract: Gaussian Process Networks (GPNs) are a class of directed graphical models which employ Gaussian processes as priors for the conditional expectation of each variable given its parents in the network. The model allows the description of continuous joint distributions in a compact but flexible manner with minimal parametric assumptions on the dependencies between variables. Bayesian structure learning of GPNs requires computing the posterior over graphs of the network and is computationally infeasible even in low dimensions. This work implements Monte Carlo and Markov Chain Monte Carlo methods to sample from the posterior distribution of network structures. As such, the approach follows the Bayesian paradigm, comparing models via their marginal likelihood and computing the posterior probability of the GPN features. Simulation studies show that our method outperforms state-of-the-art algorithms in recovering the graphical structure of the network and provides an accurate approximation of its posterior distribution.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (52)
  1. Discovering causal structures in Bayesian Gaussian directed acyclic graph models. Journal of the Royal Statistical Society: Series A (Statistics in Society), 183:1727–1745, 2020.
  2. Learning Bayesian networks: approaches and issues. The Knowledge Engineering Review, 26:99–157, 2011.
  3. Luis M. de Campos. A scoring function for learning Bayesian networks based on mutual information and conditional independence tests. Journal of Machine Learning Research, 7:2149–2187, 2006.
  4. Structure discovery in nonparametric regression through compositional kernel search. In Proceedings of the 30th International Conference on Machine Learning, volume 28, pages 1166–1174, 2013.
  5. Additive Gaussian processes. In Advances in Neural Information Processing Systems, volume 24, pages 226–234, 2011.
  6. Gal Elidan. Lightning-speed structure learning of nonlinear continuous networks. In Proceedings of the Fifteenth International Conference on Artificial Intelligence and Statistics, volume 22, pages 355–363, 2012.
  7. Learning causal Bayesian network structures from experimental data. Journal of the American Statistical Association, 103:778–789, 2008.
  8. Being Bayesian about network structure: a Bayesian approach to structure discovery in Bayesian networks. Machine Learning, 50:95–125, 2003.
  9. Gaussian process networks. In Proceedings of the Sixteenth Conference on Uncertainty in Artificial Intelligence, pages 211–219, 2000.
  10. Data analysis with Bayesian networks: a bootstrap approach. In Proceedings of the Fifteenth Conference on Uncertainty in Artificial Intelligence, pages 196–205, 2013.
  11. Learning Gaussian networks. In Proceedings of the Tenth Conference on Uncertainty in Artificial Intelligence, pages 235–243, 1994.
  12. Parameter priors for directed acyclic graphical models and the characterization of several probability distributions. The Annals of Statistics, 30:1412–1440, 2002.
  13. The dual PC algorithm and the role of Gaussianity for structure learning of Bayesian networks. International Journal of Approximate Reasoning, 161:108975, 2023.
  14. Improving Markov chain Monte Carlo model search for data mining. Machine Learning, 50:127–158, 2003.
  15. Robert B. Gramacy and Herbert K. H. Lee. Bayesian treed Gaussian process models with an application to computer modeling. Journal of the American Statistical Association, 103:1119–1130, 2008.
  16. A kernel statistical test of independence. In Advances in Neural Information Processing Systems, volume 20, pages 585–592, 2007.
  17. Nonlinear directed acyclic structure learning with weakly additive noise models. In Advances in Neural Information Processing Systems, volume 22, pages 1847–1855, 2009.
  18. A tutorial on bridge sampling. Journal of Mathematical Psychology, 81:80–97, 2017.
  19. Bridgesampling: An R package for estimating normalizing constants. Journal of Statistical Software, 92:1–29, 2020.
  20. Improving the structure MCMC sampler for Bayesian networks by introducing a new edge reversal move. Machine Learning, 71:265–305, 2008.
  21. Learning Bayesian networks: a unification for discrete and Gaussian domains. Proceedings of the Eleventh Conference on Uncertainty in Artificial Intelligence, pages 274–284, 1995.
  22. The No-U-turn sampler: adaptively setting path lengths in Hamiltonian Monte Carlo. Journal of Machine Learning Research, 15:1593–1623, 2014.
  23. A survey of Bayesian network structure learning. Artificial Intelligence Review, 56:1–94, 2023.
  24. Probabilistic graphical models: principles and techniques - adaptive computation and machine learning. The MIT Press, 2009.
  25. Partition MCMC for inference on acyclic digraphs. Journal of the American Statistical Association, 112:282–299, 2017.
  26. Addendum on the scoring of Gaussian directed acyclic graphical models. The Annals of Statistics, 42:1689–1691, 2014.
  27. Mutational interactions define novel cancer subgroups. Nature Communications, 9:4353, 2018.
  28. Efficient sampling and structure learning of Bayesian networks. Journal of Computational and Graphical Statistics, 31:639–650, 2022.
  29. DiBS: differentiable Bayesian structure learning. In Advances in Neural Information Processing Systems, volume 34, pages 24111–24123, 2021.
  30. Bayesian graphical models for discrete data. International Statistical Review, 63:215–232, 1995.
  31. Simulating ratios of normalizing constants via a simple identity: a theoretical exploration. Statistica Sinica, 6:831–860, 1996.
  32. Longitudinal symptomatic interactions in long-standing schizophrenia: a novel five-point analysis based on directed acyclic graphs. Psychological Medicine, pages 1–8, 2021.
  33. Judea Pearl. Probabilistic reasoning in intelligent systems: networks of plausible inference. Morgan Kaufmann Publishers Inc., 1988.
  34. Judea Pearl. Causality: Models, reasoning, and inference. Cambridge University Press, 2000.
  35. Causal discovery with continuous additive noise models. Journal of Machine Learning Research, 15:2009–2053, 2013.
  36. Carl Edward Rasmussen. Gaussian processes in machine learning. Springer Berlin Heidelberg, 2004.
  37. Benchpress: a scalable and platform-independent workflow for benchmarking structure learning algorithms for graphical models. arXiv:2107.03863, 2021.
  38. Robert W. Robinson. Counting labeled acyclic digraphs. In New Directions in Graph Theory, pages 239–273. New York: Academic Press, 1973.
  39. Causal protein-signaling networks derived from multiparameter single-cell data. Science, 308:523–529, 2005.
  40. Charupriya Sharma and Peter van Beek. Scalable Bayesian network structure learning with splines. In Proceedings of The 11th International Conference on Probabilistic Graphical Models, volume 186, pages 181–192, 2022.
  41. Locally minimax optimal predictive modeling with Bayesian networks. In Proceedings of the Twelth International Conference on Artificial Intelligence and Statistics, volume 5, pages 504–511, 2009.
  42. Causation, prediction, and search. Springer New York, NY, 1993.
  43. Stan Development Team. RStan: the R interface to Stan, R package version 2.21.8, 2023. URL https://mc-stan.org/.
  44. Improving structure MCMC for Bayesian networks through Markov blanket resampling. Journal of Machine Learning Research, 17:1–20, 2016.
  45. Bayesian structure learning and sampling of Bayesian networks with the R package BiDAG. Journal of Statistical Software, 105:1–31, 2023.
  46. Measuring and testing dependence by correlation of distances. The Annals of Statistics, 35:2769–2794, 2007.
  47. Exact sampling of directed acyclic graphs from modular distributions. In Proceedings of The 35th Uncertainty in Artificial Intelligence Conference, pages 965–974, 2020.
  48. The max-min hill-climbing Bayesian network structure learning algorithm. Machine Learning, 65:31–78, 2006.
  49. Towards scalable Bayesian learning of causal DAGs. In Advances in Neural Information Processing Systems, volume 33, pages 6584–6594, 2020.
  50. Properties of the bridge sampler with a focus on splitting the MCMC sample. Statistics and Computing, 30:799–816, 2020.
  51. DAG-GNN: DAG structure learning with graph neural networks. In Proceedings of the 36th International Conference on Machine Learning, volume 97, pages 7154–7163, 2019.
  52. DAGs with NO TEARS: continuous optimization for structure learning. In Advances in Neural Information Processing Systems, volume 31, pages 9492–9503, 2018.
Citations (3)

Summary

We haven't generated a summary for this paper yet.