A Compositional Framework for First-Order Optimization (2403.05711v1)
Abstract: Optimization decomposition methods are a fundamental tool to develop distributed solution algorithms for large scale optimization problems arising in fields such as machine learning and optimal control. In this paper, we present an algebraic framework for hierarchically composing optimization problems defined on hypergraphs and automatically generating distributed solution algorithms that respect the given hierarchical structure. The central abstractions of our framework are operads, operad algebras, and algebra morphisms, which formalize notions of syntax, semantics, and structure preserving semantic transformations respectively. These abstractions allow us to formally relate composite optimization problems to the distributed algorithms that solve them. Specifically, we show that certain classes of optimization problems form operad algebras, and a collection of first-order solution methods, namely gradient descent, Uzawa's algorithm (also called gradient ascent-descent), and their subgradient variants, yield algebra morphisms from these problem algebras to algebras of dynamical systems. Primal and dual decomposition methods are then recovered by applying these morphisms to certain classes of composite problems. Using this framework, we also derive a novel sufficient condition for when a problem defined by compositional data is solvable by a decomposition method. We show that the minimum cost network flow problem satisfies this condition, thereby allowing us to automatically derive a hierarchical dual decomposition algorithm for finding minimum cost flows on composite flow networks. We implement our operads, algebras, and algebra morphisms in a Julia package called AlgebraicOptimization.jl and use our implementation to empirically demonstrate that hierarchical dual decomposition outperforms standard dual decomposition on classes of flow networks with hierarchical structure.
- Stephen Boyd. Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers. Foundations and Trends® in Machine Learning, 3(1):1–122, 2010.
- Block splitting for distributed optimization. Mathematical Programming Computation, 6(1):77–102, March 2014.
- Accelerated gradient methods and dual decomposition in distributed model predictive control. Automatica, 49(3):829–833, March 2013.
- A primal decomposition algorithm for distributed multistage scenario model predictive control. Journal of Process Control, 81:162–171, 2019.
- J. F. Benders. Partitioning procedures for solving mixed-variables programming problems. Numer. Math., 4(1):238–252, dec 1962.
- The Benders decomposition algorithm: A literature review. European Journal of Operational Research, 259(3):801–817, June 2017.
- Notes on decomposition methods. Notes for EE364B, Stanford University, 2015.
- SnapVX: A Network-Based Convex Optimization Solver. Journal of Machine Learning Research, 18:1–5, 2017.
- A graph-based modeling abstraction for optimization: Concepts and implementation in plasmo.jl. Mathematical Programming Computation, 14:699 – 747, 2022.
- David I. Spivak. The operad of wiring diagrams: formalizing a graphical language for databases, recursion, and plug-and-play circuits, May 2013. arXiv:1305.0297 [cs, math].
- Structured versus Decorated Cospans. Compositionality, 4:3, September 2022. arXiv:2101.09363 [math].
- A compositional framework for passive linear networks. Theory and Applications of Categories, 33(38):1158–1222, 2018.
- A Compositional Framework for Reaction Networks. Reviews in Mathematical Physics, 29(09):1750028, October 2017. arXiv:1704.02051 [math-ph].
- Categories in control. Theory and Applications of Categories, 30(24):836–881, 2015.
- Towards a compositional framework for convex analysis (with applications to probability theory), 2023.
- A compositional framework for markov processes. Journal of Mathematical Physics, 57(3), March 2016.
- Operadic Modeling of Dynamical Systems: Mathematics and Computation. Electronic Proceedings in Theoretical Computer Science, 372:192–206, November 2022. arXiv:2105.12282 [math].
- Riccardo Scattolini. Architectures for distributed and hierarchical model predictive control–a review. Journal of process control, 19(5):723–731, 2009.
- Theory of hierarchical, multilevel, systems. Elsevier, 2000.
- Control and coordination in hierarchical systems. John Wiley & Sons, 1980.
- Studies in Linear and Non-linear Programming. Standford University Press, 1958.
- Yuri Nesterov. Primal-dual subgradient methods for convex problems. Mathematical Programming, 120:221–259, 2009.
- Disciplined Convex Programming. In Leo Liberti and Nelson Maculan, editors, Global Optimization, volume 84, pages 155–210. Kluwer Academic Publishers, Boston, 2006. Series Title: Nonconvex Optimization and Its Applications.
- Backprop as functor: A compositional perspective on supervised learning. In 2019 34th Annual ACM/IEEE Symposium on Logic in Computer Science (LICS), pages 1–13. IEEE, 2019.
- Categorical foundations of gradient-based learning, 2021.
- Characterizing compositionality of lqr from the categorical perspective. Proceedings of the 62nd IEEE Conference on Decision and Control, 2023.
- Compositional game theory. In Proceedings of the 33rd annual ACM/IEEE symposium on logic in computer science, pages 472–481, 2018.
- Compositional game theory, compositionally. arXiv preprint arXiv:2101.12045, 2021.
- Seven sketches in compositionality: An invitation to applied category theory, 2018.
- E. Riehl. Category theory in context. Aurora: Dover modern math originals. Dover Publications, 2017.
- Steve Awodey. Category theory. Oxford university press, 2010.
- Physics, topology, logic and computation: a Rosetta Stone. Springer, 2011.
- R. Tyrrell Rockafellar. Convex Analysis. Princeton University Press, 1970.
- Convex Optimization. Cambridge University Press, March 2004.
- Dan Shiebler. Generalized Optimization: A First Step Towards Category Theoretic Learning Theory. In Pandian Vasant, Ivan Zelinka, and Gerhard-Wilhelm Weber, editors, Intelligent Computing & Optimization, pages 525–535, Cham, 2022. Springer International Publishing.
- Hypergraph Categories, January 2019. arXiv:1806.08304 [cs, math].
- Claudio Hermida. Representable Multicategories. Advances in Mathematics, 151(2):164–225, 2000.
- Brendan Fong. The Algebra of Open and Interconnected Systems, September 2016. arXiv:1609.05382 [math].
- Disciplined Saddle Programming, January 2023. arXiv:2301.13427 [cs, math].
- Subgradients. Notes for EE364B, Stanford University, 2022.
- Linear programming and network flows. John Wiley & Sons, 2011.
- Lester Randolph Ford. Flows in networks. Princeton University Press, 2015.
- Network flow programming. R.E. Krieger Publishing Company, 1987.
- Algorithms for network programming. John Wiley & Sons, Inc., 1980.
- James Evans. Optimization algorithms for networks and graphs. CRC Press, 2017.
- Network flows: Theory, applications and algorithms. Prentice-Hall, Englewood Cliffs, New Jersey, USA Arrow, KJ (1963). Social Choice and Individual Values, Wiley, New York. Gibbard, A.(1973).“Manipulation of Voting Schemes: A general result”, Econometrica, 41:587–602, 1993.
- David P Williamson. Network flow algorithms. Cambridge University Press, 2019.
- A distributed newton method for network optimization. In Proceedings of the 48h IEEE Conference on Decision and Control (CDC) held jointly with 2009 28th Chinese Control Conference, pages 2736–2741, Shanghai, China, December 2009. IEEE.
- Compositional scientific computing with catlab and semanticmodels. arXiv preprint arXiv:2005.04831, 2020.
- Forward-mode automatic differentiation in Julia. arXiv:1607.07892 [cs.MS], 2016.
- On random graphs. i. Publicationes Mathematicae Debrecen, 1959.
- Optim: A mathematical optimization package for Julia. Journal of Open Source Software, 3(24):615, 2018.
- Compositional thermostatics. Journal of Mathematical Physics, 64(2), February 2023.
- Categorical foundations of gradient-based learning. In European Symposium on Programming, pages 1–28. Springer International Publishing Cham, 2022.
- Totally asynchronous primal-dual convex optimization in blocks, 2022.
Sponsored by Paperpile, the PDF & BibTeX manager trusted by top AI labs.
Get 30 days freePaper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.