Papers
Topics
Authors
Recent
Detailed Answer
Quick Answer
Concise responses based on abstracts only
Detailed Answer
Well-researched responses based on abstracts and relevant paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses
Gemini 2.5 Flash
Gemini 2.5 Flash 87 tok/s
Gemini 2.5 Pro 53 tok/s Pro
GPT-5 Medium 17 tok/s Pro
GPT-5 High 20 tok/s Pro
GPT-4o 106 tok/s Pro
Kimi K2 156 tok/s Pro
GPT OSS 120B 467 tok/s Pro
Claude Sonnet 4 37 tok/s Pro
2000 character limit reached

Learning to Solve the Constrained Most Probable Explanation Task in Probabilistic Graphical Models (2404.11606v1)

Published 17 Apr 2024 in cs.LG and cs.AI

Abstract: We propose a self-supervised learning approach for solving the following constrained optimization task in log-linear models or Markov networks. Let $f$ and $g$ be two log-linear models defined over the sets $\mathbf{X}$ and $\mathbf{Y}$ of random variables respectively. Given an assignment $\mathbf{x}$ to all variables in $\mathbf{X}$ (evidence) and a real number $q$, the constrained most-probable explanation (CMPE) task seeks to find an assignment $\mathbf{y}$ to all variables in $\mathbf{Y}$ such that $f(\mathbf{x}, \mathbf{y})$ is maximized and $g(\mathbf{x}, \mathbf{y})\leq q$. In our proposed self-supervised approach, given assignments $\mathbf{x}$ to $\mathbf{X}$ (data), we train a deep neural network that learns to output near-optimal solutions to the CMPE problem without requiring access to any pre-computed solutions. The key idea in our approach is to use first principles and approximate inference methods for CMPE to derive novel loss functions that seek to push infeasible solutions towards feasible ones and feasible solutions towards optimal ones. We analyze the properties of our proposed method and experimentally demonstrate its efficacy on several benchmark problems.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (47)
  1. Achterberg, T. (2009). Scip: solving constraint integer programs. Mathematical Programming Computation, 1:1–41.
  2. Constraint integer programming: A new approach to integrate cp and mip. In Integration of AI and OR Techniques in Constraint Programming for Combinatorial Optimization Problems: 5th International Conference, CPAIOR 2008 Paris, France, May 20-23, 2008 Proceedings 5, pages 6–20. Springer.
  3. Relax, compensate and then recover. In Onada, T., Bekki, D., and McCready, E., editors, New Frontiers in Artificial Intelligence, pages 167–180, Berlin, Heidelberg. Springer Berlin Heidelberg.
  4. Same-decision probability: A confidence measure for threshold-based decisions. International Journal of Approximate Reasoning, 53(9):1415–1428.
  5. Probabilistic circuits: A unifying framework for tractable probabilistic models. UCLA. URL: http://starai. cs. ucla. edu/papers/ProbCirc20. pdf.
  6. Variational message passing neural network for maximum-a-posteriori (map) inference. In Uncertainty in Artificial Intelligence, pages 464–474. PMLR.
  7. Darwiche, A. (2009). Modeling and Reasoning with Bayesian Networks. Cambridge University Press.
  8. On the reasons behind decisions. In Twenty Fourth European Conference on Artificial Intelligence, volume 325 of Frontiers in Artificial Intelligence and Applications, pages 712–720. IOS Press.
  9. On the (complete) reasons behind decisions. Journal of Logic, Language and Information, 32(1):63–88.
  10. Dechter, R. (1999). Bucket elimination: A unifying framework for reasoning. Artificial Intelligence, 113:41–85.
  11. Mini-buckets: A general scheme for bounded inference. Journal of the ACM (JACM), 50(2):107–153.
  12. Dc3: A learning method for optimization with hard constraints. arXiv preprint arXiv:2104.12225.
  13. The 2010 UAI Approximate Inference Challenge. Published: Available online at: http://www.cs.huji.ac.il/project/UAI10/index.php.
  14. Predicting ac optimal power flows: Combining deep learning and lagrangian dual methods. In Proceedings of the AAAI conference on artificial intelligence, volume 34, pages 630–637.
  15. Neural message passing for quantum chemistry. In International conference on machine learning, pages 1263–1272. PMLR.
  16. Fixing max-product: Convergent message passing algorithms for map lp-relaxations. Advances in neural information processing systems, 20:553–560.
  17. Gogate, V. (2014). Results of the 2014 UAI competition. https://personal.utdallas.edu/~vibhav.gogate/uai14-competition/index.html.
  18. Gogate, V. (2016). Results of the 2016 UAI competition. https://personal.utdallas.edu/~vibhav.gogate/uai16-competition/index.html.
  19. Gurobi Optimization, L. (2021). Gurobi optimizer reference manual.
  20. Global Optimization: Deterministic Approaches. Springer Berlin Heidelberg.
  21. Join-graph based cost-shifting schemes. arXiv preprint arXiv:1210.4878.
  22. Adam: A Method for Stochastic Optimization. arXiv.
  23. Probabilistic graphical models: principles and techniques. MIT press.
  24. Mrf optimization via dual decomposition: Message-passing revisited. In 2007 IEEE 11th International Conference on Computer Vision, pages 1–8. IEEE.
  25. Learning hard optimization problems: A data generation perspective. arXiv preprint arXiv: Arxiv-2106.02601.
  26. Belief propagation neural networks. Advances in Neural Information Processing Systems, 33:667–678.
  27. MNIST handwritten digit database.
  28. Learning a constrained optimizer: A primal method. In AAAI 2023 Bridge on Constraint Programming and Machine Learning.
  29. Learning Markov Network Structure with Decision Trees. In 2010 IEEE International Conference on Data Mining, pages 334–343. IEEE.
  30. Memory intensive AND/OR search for combinatorial optimization in graphical models. AI Journal, 173(16-17):1492–1524.
  31. Best-First AND/OR Search for Most Probable Explanations. CoRR, abs/1206.5268.
  32. Physics-informed neural networks for ac optimal power flow. arXiv preprint arXiv: Arxiv-2110.02672.
  33. Numerical Optimization. Springer, New York, NY, USA, 2e edition.
  34. Self-supervised primal-dual learning for constrained optimization. arXiv preprint arXiv:2208.09046.
  35. Cutset networks: A simple, tractable, and scalable approach for improving the accuracy of chow-liu trees. In Machine Learning and Knowledge Discovery in Databases: European Conference, ECML PKDD 2014, Nancy, France, September 15-19, 2014. Proceedings, Part II 14, pages 630–645. Springer.
  36. Novel upper bounds for the constrained most probable explanation task. Advances in Neural Information Processing Systems, 34:9613–9624.
  37. Algorithms for the nearest assignment problem. In Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, IJCAI 2018, July 13-19, 2018, Stockholm, Sweden, pages 5096–5102. ijcai.org.
  38. A novel approach for constrained optimization in graphical models. Advances in Neural Information Processing Systems, 33:11949–11960.
  39. Neural enhanced belief propagation on factor graphs. In International Conference on Artificial Intelligence and Statistics, pages 685–693. PMLR.
  40. A reformulation-linearization technique (rlt) for semi-infinite and convex programs under mixed 0-1 and general discrete restrictions. Discrete Applied Mathematics, 157(6):1319–1333. Reformulation Techniques and Mathematical Programming.
  41. A global optimization algorithm for polynomial programming problems using a reformulation-linearization technique. Journal of Global Optimization, 2:101–112.
  42. Ucla-Starai (2023). Density-Estimation-Datasets. [Online; accessed 17. May 2023].
  43. Markov Network Structure Learning: A Randomized Feature Generation Approach. AAAI, 26(1):1148–1154.
  44. Map estimation via agreement on trees: message-passing and linear programming. IEEE transactions on information theory, 51(11):3697–3717.
  45. Map inference via l2 sphere linear program reformulation. International Journal of Computer Vision, 128(7):1913–1936.
  46. Learning optimal solutions for extremely fast ac optimal power flow. arXiv preprint arXiv: Arxiv-1910.01213.
  47. Factor graph neural networks. Advances in Neural Information Processing Systems, 33:8577–8587.
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-Up Questions

We haven't generated follow-up questions for this paper yet.