Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
125 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Neural Network Approximators for Marginal MAP in Probabilistic Circuits (2402.03621v1)

Published 6 Feb 2024 in cs.LG and cs.AI

Abstract: Probabilistic circuits (PCs) such as sum-product networks efficiently represent large multi-variate probability distributions. They are preferred in practice over other probabilistic representations such as Bayesian and Markov networks because PCs can solve marginal inference (MAR) tasks in time that scales linearly in the size of the network. Unfortunately, the maximum-a-posteriori (MAP) and marginal MAP (MMAP) tasks remain NP-hard in these models. Inspired by the recent work on using neural networks for generating near-optimal solutions to optimization problems such as integer linear programming, we propose an approach that uses neural networks to approximate (M)MAP inference in PCs. The key idea in our approach is to approximate the cost of an assignment to the query variables using a continuous multilinear function, and then use the latter as a loss function. The two main benefits of our new method are that it is self-supervised and after the neural network is learned, it requires only linear time to output a solution. We evaluate our new approach on several benchmark datasets and show that it outperforms three competing linear time approximations, max-product inference, max-marginal inference and sequential estimation, which are used in practice to solve MMAP tasks in PCs.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (41)
  1. Probabilistic circuits: A unifying framework for tractable probabilistic models. Technical report, University of California, Los Angeles, 2020.
  2. Sum-Product Networks: A New Deep Architecture. In Proceedings of the Twenty-Seventh Conference on Uncertainty in Artificial Intelligence, pages 337–346. AUAI Press, 2011.
  3. Adnan Darwiche. A differential approach to inference in bayesian networks. Journal of the ACM (JACM), 50(3):280–305, 2003.
  4. And/or search spaces for graphical models. Artificial intelligence, 171(2-3):73–106, 2007.
  5. Cutset networks: A simple, tractable, and scalable approach for improving the accuracy of chow-liu trees. In Machine Learning and Knowledge Discovery in Databases: European Conference, ECML PKDD 2014, Nancy, France, September 15-19, 2014. Proceedings, Part II 14, pages 630–645. Springer, 2014.
  6. Probabilistic sentential decision diagrams. In Fourteenth International Conference on the Principles of Knowledge Representation and Reasoning, 2014.
  7. A compositional atlas of tractable circuit operations for probabilistic inference. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, volume 34, pages 13189–13201. Curran Associates, Inc., 2021. URL https://proceedings.neurips.cc/paper_files/paper/2021/file/6e01383fd96a17ae51cc3e15447e7533-Paper.pdf.
  8. Robert Peharz. Foundations of sum-product networks for probabilistic modeling. PhD thesis, PhD thesis, Medical University of Graz, 2015.
  9. On the latent variable interpretation in sum-product networks. IEEE transactions on pattern analysis and machine intelligence, 39(10):2030–2044, 2016.
  10. Bayesian image segmentation using hidden fields: Supervised, unsupervised, and semi-supervised formulations. In 2016 24th European Signal Processing Conference (EUSIPCO), pages 523–527, August 2016. doi:10.1109/EUSIPCO.2016.7760303.
  11. POMDP Planning by Marginal-MAP Probabilistic Inference in Generative Models. In Proceedings of the 2014 AAMAS Workshop on Adaptive Learning Agents, 2014.
  12. Applying marginal MAP search to probabilistic conformant planning: Initial results. In Statistical Relational Artificial Intelligence, Papers from the 2014 AAAI Workshop, Québec City, Québec, Canada, July 27, 2014, volume WS-14-13 of AAAI Technical Report. AAAI, 2014. URL http://www.aaai.org/ocs/index.php/WS/AAAIW14/paper/view/8855.
  13. Decomposition bounds for marginal map. In C. Cortes, N. Lawrence, D. Lee, M. Sugiyama, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 28. Curran Associates, Inc., 2015. URL https://proceedings.neurips.cc/paper_files/paper/2015/file/faacbcd5bf1d018912c116bf2783e9a1-Paper.pdf.
  14. Approximation complexity of maximum A posteriori inference in sum-product networks. In Gal Elidan, Kristian Kersting, and Alexander Ihler, editors, Proceedings of the Thirty-Third Conference on Uncertainty in Artificial Intelligence (UAI). AUAI Press, 2017.
  15. Maximum a posteriori inference in sum-product networks. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 32, 2018.
  16. Complexity results and approximation strategies for map explanations. J. Artif. Int. Res., 21(1):101–133, feb 2004. ISSN 1076-9757.
  17. Cassio P de Campos. New Complexity Results for MAP in Bayesian Networks. IJCAI International Joint Conference on Artificial Intelligence, pages 2100–2106, 01 2011. doi:10.5591/978-1-57735-516-8/IJCAI11-351.
  18. Two reformulation approaches to maximum-a-posteriori inference in sum-product networks. In International Conference on Probabilistic Graphical Models, pages 293–304. PMLR, 2020.
  19. Solving marginal map exactly by probabilistic circuit transformations. In International Conference on Artificial Intelligence and Statistics, pages 10196–10208. PMLR, 2022.
  20. Look ma, no latent variables: Accurate cutset networks via compilation. In International Conference on Machine Learning, pages 5311–5320. PMLR, 2019.
  21. Ke Li and Jitendra Malik. Learning to optimize, 2016.
  22. Predicting AC Optimal Power Flows: Combining Deep Learning and Lagrangian Dual Methods. Proceedings of the AAAI Conference on Artificial Intelligence, 34(01):630–637, April 2020. ISSN 2374-3468. doi:10.1609/aaai.v34i01.5403.
  23. DC3: A learning method for optimization with hard constraints. In International Conference on Learning Representations, October 2020.
  24. Learning Optimal Solutions for Extremely Fast AC Optimal Power Flow. In 2020 IEEE International Conference on Communications, Control, and Computing Technologies for Smart Grids (SmartGridComm), pages 1–6, November 2020. doi:10.1109/SmartGridComm47815.2020.9303008.
  25. Self-Supervised Primal-Dual Learning for Constrained Optimization. Proceedings of the AAAI Conference on Artificial Intelligence, 37(4):4052–4060, June 2023. ISSN 2374-3468. doi:10.1609/aaai.v37i4.25520.
  26. Inference in probabilistic graphical models by graph neural networks. In 2019 53rd Asilomar Conference on Signals, Systems, and Computers, pages 868–875. IEEE, 2019.
  27. Variational message passing neural network for Maximum-A-Posteriori (MAP) inference. In James Cussens and Kun Zhang, editors, Uncertainty in Artificial Intelligence, Proceedings of the Thirty-Eighth Conference on Uncertainty in Artificial Intelligence, UAI 2022, 1-5 August 2022, Eindhoven, the Netherlands, volume 180 of Proceedings of Machine Learning Research, pages 464–474. PMLR, 2022.
  28. T. Rahman and V. Gogate. Merging strategies for sum-product networks: From trees to graphs. In Proceedings of the Thirty-Second Conference Conference on Uncertainty in Artificial Intelligence, pages 617–626, 2016a.
  29. Learning Ensembles of Cutset Networks. Proceedings of the AAAI Conference on Artificial Intelligence, 30(1), March 2016b. ISSN 2374-3468. doi:10.1609/aaai.v30i1.10428.
  30. V. Gogate and R. Dechter. Importance sampling-based estimation over AND/OR search spaces for graphical models. Artificial Intelligence, 184-185:38–77, 2012.
  31. Learning Markov Network Structure with Decision Trees. In 2010 IEEE International Conference on Data Mining, pages 334–343. IEEE, 2010. ISBN 978-1-4244-9131-5. doi:10.1109/ICDM.2010.128. URL http://ieeexplore.ieee.org/document/5693987/.
  32. Markov Network Structure Learning: A Randomized Feature Generation Approach. Proceedings of the AAAI Conference on Artificial Intelligence, 26(1):1148–1154, 2012. ISSN 2374-3468. doi:10.1609/aaai.v26i1.8315. URL https://ojs.aaai.org/index.php/AAAI/article/view/8315.
  33. The neural autoregressive distribution estimator. In Geoffrey Gordon, David Dunson, and Miroslav Dudík, editors, Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics, volume 15 of Proceedings of Machine Learning Research, pages 29–37, Fort Lauderdale, FL, USA, 11–13 Apr 2011. PMLR. URL https://proceedings.mlr.press/v15/larochelle11a.html.
  34. Tractable learning for complex probability queries. In C. Cortes, N. Lawrence, D. Lee, M. Sugiyama, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 28. Curran Associates, Inc., 2015.
  35. On the quantitative analysis of deep belief networks. In Proceedings of the 25th international conference on Machine learning, pages 872–879. ACM, 2008.
  36. EMNIST: An extension of MNIST to handwritten letters, February 2017.
  37. Learning multiple layers of features from tiny images. Technical report, Technical Report, University of Toronto, 2009.
  38. DeeProb-kit: a python library for deep probabilistic modelling, 2022. URL https://arxiv.org/abs/2212.04403.
  39. Dropout: A Simple Way to Prevent Neural Networks from Overfitting. Journal of Machine Learning Research, 15(56):1929–1958, 2014. ISSN 1533-7928.
  40. Adam: A Method for Stochastic Optimization, January 2017.
  41. Pytorch: An imperative style, high-performance deep learning library. Advances in neural information processing systems, 32, 2019.

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com