Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
156 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Convergence of Some Convex Message Passing Algorithms to a Fixed Point (2403.07004v2)

Published 7 Mar 2024 in cs.AI, cs.LG, math.OC, and stat.ML

Abstract: A popular approach to the MAP inference problem in graphical models is to minimize an upper bound obtained from a dual linear programming or Lagrangian relaxation by (block-)coordinate descent. This is also known as convex/convergent message passing; examples are max-sum diffusion and sequential tree-reweighted message passing (TRW-S). Convergence properties of these methods are currently not fully understood. They have been proved to converge to the set characterized by local consistency of active constraints, with unknown convergence rate; however, it was not clear if the iterates converge at all (to any point). We prove a stronger result (conjectured before but never proved): the iterates converge to a fixed point of the method. Moreover, we show that the algorithm terminates within $\mathcal{O}(1/\varepsilon)$ iterations. We first prove this for a version of coordinate descent applied to a general piecewise-affine convex objective. Then we show that several convex message passing methods are special cases of this method. Finally, we show that a slightly different version of coordinate descent can cycle.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (34)
  1. FastDOG: Fast discrete optimization on GPU. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  439–449, 2022.
  2. Bertsekas, D. P. Nonlinear Programming. Athena Scientific, Belmont, MA, 2nd edition, 1999.
  3. Learning deep structured models. In International Conference on Machine Learning, pp.  1785–1794. PMLR, 2015.
  4. Soft arc consistency revisited. Artificial Intelligence, 174(7-8):449–478, 2010.
  5. Fixing max-product: Convergent message passing algorithms for MAP LP-relaxations. In Neural Information Processing Systems, pp.  553–560, 2008.
  6. Lagrangean decomposition: A model yielding stronger Lagrangean bounds. Mathematical Programming, 39:215––228, 1987.
  7. Lagrangian relaxation for MAP estimation in graphical models. In 45th Allerton Conference on Communication, Control and Computing, 2007.
  8. A comparative study of modern inference techniques for structured discrete energy minimization problems. Intl. J. of Computer Vision, 115(2):155–184, 2015. ISSN 1573-1405.
  9. Kolmogorov, V. Convergent tree-reweighted message passing for energy minimization. IEEE Trans. Pattern Analysis and Machine Intelligence, 28(10):1568–1583, 2006.
  10. Kolmogorov, V. A new look at reweighted message passing. IEEE Trans. on Pattern Analysis and Machine Intelligence, 37(5), May 2015.
  11. MRF energy minimization and beyond via dual decomposition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 33(3):531–552, 2011.
  12. A diffusion algorithm for decreasing the energy of the max-sum labeling problem. Glushkov Institute of Cybernetics, Kiev, USSR. Unpublished, 1975.
  13. Efficient message passing for 0-1 ILPs with binary decision diagrams. In Intl. Conf. on Machine Learning (ICML), volume 139, pp.  6000–6010. PMLR, 2021.
  14. An augmented Lagrangian approach to constrained MAP inference. In Intl. Conf. on Machine Learning, pp.  169–176, 2011.
  15. Soft constraints. In Handbook of Constraint Programming, chapter 9. Elsevier, 2006.
  16. Scalable full flow with learned binary descriptors. In Pattern Recognition: 39th German Conference, GCPR 2017, Basel, Switzerland, September 12–15, 2017, Proceedings 39, pp.  321–332. Springer, 2017.
  17. Savchynskyy, B. A bundle approach to efficient MAP-inference by Lagrangian relaxation. In IEEE Conf. on Computer Vision and Pattern Recognition, pp.  1688–1695, 2012.
  18. Savchynskyy, B. Discrete graphical models – an optimization perspective. Foundations and Trends in Computer Graphics and Vision, 11(3-4):160–429, 2019. ISSN 1572-2740.
  19. Efficient MRF energy minimization via adaptive diminishing smoothing. In Conf. on Uncertainty in Artificial Intelligence, pp.  746–755, 2012.
  20. Diffusion algorithms and structural recognition optimization problems. Cybernetics and Systems Analysis, 47:175–192, 2011. ISSN 1060-0396.
  21. Solving dense image matching in real-time using discrete-continuous optimization. arXiv preprint arXiv:1601.06274, 2016.
  22. Introduction to dual decomposition for inference. In Sra, S., Nowozin, S., and Wright, S. J. (eds.), Optimization for Machine Learning. MIT Press, 2012.
  23. A message passing algorithm for the minimum cost multicut problem. In Conf. on Computer Vision and Pattern Recognition, 2017.
  24. A dual ascent framework for Lagrangean decomposition of combinatorial problems. In Conf. on Computer Vision and Pattern Recognition, 2017a.
  25. A study of Lagrangean decompositions and dual ascent solvers for graph matching. In Conf. on Computer Vision and Pattern Recognition, 2017b.
  26. A comparative study of energy minimization methods for Markov random fields. In Eur. Conf. on Computer Vision, pp.  II: 16–29, 2006.
  27. Mplp++: Fast, parallel dual block-coordinate ascent for dense graphical models. In Proceedings of the European Conference on Computer Vision (ECCV), pp.  251–267, 2018.
  28. Taxonomy of dual block-coordinate ascent methods for discrete energy minimization. In Conference on Artificial Intelligence and Statistics, Proceedings of Machine Learning Research, pp.  2775–2785. PMLR, 2020.
  29. Tseng, P. Convergence of a block coordinate descent method for nondifferentiable minimization. J. Optim. Theory Appl., 109(3):475–494, June 2001.
  30. Graphical models, exponential families, and variational inference. Foundations and Trends in Machine Learning, 1(1-2):1–305, 2008.
  31. Wedelin, D. An algorithm for large scale 0-1 integer programming with application to airline crew scheduling. Annals of Operation Research, 57(1):283–301, 1995.
  32. Werner, T. A linear programming approach to max-sum problem: A review. IEEE Trans. Pattern Analysis and Machine Intelligence, 29(7):1165–1179, July 2007.
  33. Werner, T. Revisiting the linear programming relaxation approach to Gibbs energy minimization and weighted constraint satisfaction. IEEE Trans. Pattern Analysis and Machine Intelligence, 32(8):1474–1488, August 2010.
  34. Relative interior rule in block-coordinate descent. In Conf. on Computer Vision and Pattern Recognition (CVPR), June 2020.

Summary

We haven't generated a summary for this paper yet.