Convergence of Some Convex Message Passing Algorithms to a Fixed Point (2403.07004v2)
Abstract: A popular approach to the MAP inference problem in graphical models is to minimize an upper bound obtained from a dual linear programming or Lagrangian relaxation by (block-)coordinate descent. This is also known as convex/convergent message passing; examples are max-sum diffusion and sequential tree-reweighted message passing (TRW-S). Convergence properties of these methods are currently not fully understood. They have been proved to converge to the set characterized by local consistency of active constraints, with unknown convergence rate; however, it was not clear if the iterates converge at all (to any point). We prove a stronger result (conjectured before but never proved): the iterates converge to a fixed point of the method. Moreover, we show that the algorithm terminates within $\mathcal{O}(1/\varepsilon)$ iterations. We first prove this for a version of coordinate descent applied to a general piecewise-affine convex objective. Then we show that several convex message passing methods are special cases of this method. Finally, we show that a slightly different version of coordinate descent can cycle.
- FastDOG: Fast discrete optimization on GPU. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 439–449, 2022.
- Bertsekas, D. P. Nonlinear Programming. Athena Scientific, Belmont, MA, 2nd edition, 1999.
- Learning deep structured models. In International Conference on Machine Learning, pp. 1785–1794. PMLR, 2015.
- Soft arc consistency revisited. Artificial Intelligence, 174(7-8):449–478, 2010.
- Fixing max-product: Convergent message passing algorithms for MAP LP-relaxations. In Neural Information Processing Systems, pp. 553–560, 2008.
- Lagrangean decomposition: A model yielding stronger Lagrangean bounds. Mathematical Programming, 39:215––228, 1987.
- Lagrangian relaxation for MAP estimation in graphical models. In 45th Allerton Conference on Communication, Control and Computing, 2007.
- A comparative study of modern inference techniques for structured discrete energy minimization problems. Intl. J. of Computer Vision, 115(2):155–184, 2015. ISSN 1573-1405.
- Kolmogorov, V. Convergent tree-reweighted message passing for energy minimization. IEEE Trans. Pattern Analysis and Machine Intelligence, 28(10):1568–1583, 2006.
- Kolmogorov, V. A new look at reweighted message passing. IEEE Trans. on Pattern Analysis and Machine Intelligence, 37(5), May 2015.
- MRF energy minimization and beyond via dual decomposition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 33(3):531–552, 2011.
- A diffusion algorithm for decreasing the energy of the max-sum labeling problem. Glushkov Institute of Cybernetics, Kiev, USSR. Unpublished, 1975.
- Efficient message passing for 0-1 ILPs with binary decision diagrams. In Intl. Conf. on Machine Learning (ICML), volume 139, pp. 6000–6010. PMLR, 2021.
- An augmented Lagrangian approach to constrained MAP inference. In Intl. Conf. on Machine Learning, pp. 169–176, 2011.
- Soft constraints. In Handbook of Constraint Programming, chapter 9. Elsevier, 2006.
- Scalable full flow with learned binary descriptors. In Pattern Recognition: 39th German Conference, GCPR 2017, Basel, Switzerland, September 12–15, 2017, Proceedings 39, pp. 321–332. Springer, 2017.
- Savchynskyy, B. A bundle approach to efficient MAP-inference by Lagrangian relaxation. In IEEE Conf. on Computer Vision and Pattern Recognition, pp. 1688–1695, 2012.
- Savchynskyy, B. Discrete graphical models – an optimization perspective. Foundations and Trends in Computer Graphics and Vision, 11(3-4):160–429, 2019. ISSN 1572-2740.
- Efficient MRF energy minimization via adaptive diminishing smoothing. In Conf. on Uncertainty in Artificial Intelligence, pp. 746–755, 2012.
- Diffusion algorithms and structural recognition optimization problems. Cybernetics and Systems Analysis, 47:175–192, 2011. ISSN 1060-0396.
- Solving dense image matching in real-time using discrete-continuous optimization. arXiv preprint arXiv:1601.06274, 2016.
- Introduction to dual decomposition for inference. In Sra, S., Nowozin, S., and Wright, S. J. (eds.), Optimization for Machine Learning. MIT Press, 2012.
- A message passing algorithm for the minimum cost multicut problem. In Conf. on Computer Vision and Pattern Recognition, 2017.
- A dual ascent framework for Lagrangean decomposition of combinatorial problems. In Conf. on Computer Vision and Pattern Recognition, 2017a.
- A study of Lagrangean decompositions and dual ascent solvers for graph matching. In Conf. on Computer Vision and Pattern Recognition, 2017b.
- A comparative study of energy minimization methods for Markov random fields. In Eur. Conf. on Computer Vision, pp. II: 16–29, 2006.
- Mplp++: Fast, parallel dual block-coordinate ascent for dense graphical models. In Proceedings of the European Conference on Computer Vision (ECCV), pp. 251–267, 2018.
- Taxonomy of dual block-coordinate ascent methods for discrete energy minimization. In Conference on Artificial Intelligence and Statistics, Proceedings of Machine Learning Research, pp. 2775–2785. PMLR, 2020.
- Tseng, P. Convergence of a block coordinate descent method for nondifferentiable minimization. J. Optim. Theory Appl., 109(3):475–494, June 2001.
- Graphical models, exponential families, and variational inference. Foundations and Trends in Machine Learning, 1(1-2):1–305, 2008.
- Wedelin, D. An algorithm for large scale 0-1 integer programming with application to airline crew scheduling. Annals of Operation Research, 57(1):283–301, 1995.
- Werner, T. A linear programming approach to max-sum problem: A review. IEEE Trans. Pattern Analysis and Machine Intelligence, 29(7):1165–1179, July 2007.
- Werner, T. Revisiting the linear programming relaxation approach to Gibbs energy minimization and weighted constraint satisfaction. IEEE Trans. Pattern Analysis and Machine Intelligence, 32(8):1474–1488, August 2010.
- Relative interior rule in block-coordinate descent. In Conf. on Computer Vision and Pattern Recognition (CVPR), June 2020.