Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
169 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Accelerating Cutting-Plane Algorithms via Reinforcement Learning Surrogates (2307.08816v2)

Published 17 Jul 2023 in cs.LG, cs.AI, and math.OC

Abstract: Discrete optimization belongs to the set of $\mathcal{NP}$-hard problems, spanning fields such as mixed-integer programming and combinatorial optimization. A current standard approach to solving convex discrete optimization problems is the use of cutting-plane algorithms, which reach optimal solutions by iteratively adding inequalities known as \textit{cuts} to refine a feasible set. Despite the existence of a number of general-purpose cut-generating algorithms, large-scale discrete optimization problems continue to suffer from intractability. In this work, we propose a method for accelerating cutting-plane algorithms via reinforcement learning. Our approach uses learned policies as surrogates for $\mathcal{NP}$-hard elements of the cut generating procedure in a way that (i) accelerates convergence, and (ii) retains guarantees of optimality. We apply our method on two types of problems where cutting-plane algorithms are commonly used: stochastic optimization, and mixed-integer quadratic programming. We observe the benefits of our method when applied to Benders decomposition (stochastic optimization) and iterative loss approximation (quadratic programming), achieving up to $45\%$ faster average convergence when compared to modern alternative algorithms.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (26)
  1. Benders Decomposition for Production Routing Under Demand Uncertainty. Operations Research, 63(4): 851–867.
  2. Stabilized Benders Methods for Large-Scale Combinatorial Optimization, with Application to Data Privacy. Management Science, 66(7): 3051–3068.
  3. Benders, J. F. 1962. Partitioning procedures for solving mixed-variables programming problems. Numerische Mathematik, 4: 238–252.
  4. Best subset selection via a modern optimization lens.
  5. An algorithmic framework for convex mixed integer nonlinear programs. Discrete optimization, 5(2): 186–204.
  6. Partial Benders decomposition strategies for two-stage stochastic integer programs, volume 37. CIRRELT.
  7. Sparse regularization for fiber ODF reconstruction: From the suboptimality of ℓℓ\ellroman_ℓ2 and ℓℓ\ellroman_ℓ1 priors to ℓℓ\ellroman_ℓ0. Medical Image Analysis, 18(6): 820–833.
  8. Reinforcement learning with combinatorial actions: An application to vehicle routing. Advances in Neural Information Processing Systems, 33: 609–620.
  9. Sparse high-dimensional models in economics. Annu. Rev. Econ., 3(1): 291–317.
  10. Cardinality-regularized hawkes-granger model. Advances in Neural Information Processing Systems, 34: 2682–2694.
  11. Jeroslow, R. C. 1973. There cannot be any algorithm for integer programming with quadratic constraints. Operations Research, 21(1): 221–224.
  12. Accelerating Generalized Benders Decomposition for Wireless Resource Allocation. IEEE Transactions on Wireless Communications, 20(2): 1233–1247.
  13. Learning Sparse Neural Networks through L_0 Regularization. In International Conference on Learning Representations.
  14. Accelerating Benders decomposition: Algorithmic enhancement and model selection criteria. Operations Research, 29(3): 464–484.
  15. Reinforcement learning for combinatorial optimization: A survey. Computers & Operations Research, 134: 105400.
  16. Solving mixed integer programs using neural networks. arXiv preprint arXiv:2012.13349.
  17. Discrete optimization. Elsevier.
  18. Improving benders decomposition using a genetic algorithm. European Journal of Operational Research, 199(1): 89–97.
  19. The Benders decomposition algorithm: A literature review. European Journal of Operational Research, 259(3): 801–817.
  20. High-Dimensional Continuous Control Using Generalized Advantage Estimation. In Proceedings of the International Conference on Learning Representations (ICLR).
  21. Proximal policy optimization algorithms. arXiv preprint arXiv:1707.06347.
  22. Reinforcement Learning: An Introduction. The MIT Press, second edition.
  23. Policy Gradient Methods for Reinforcement Learning with Function Approximation. In Solla, S.; Leen, T.; and Müller, K., eds., Advances in Neural Information Processing Systems, volume 12. MIT Press.
  24. Reinforcement learning for integer programming: Learning to cut. In International conference on machine learning, 9367–9376. PMLR.
  25. Learning Cut Selection for Mixed-Integer Linear Programming via Hierarchical Sequence Model. In The Eleventh International Conference on Learning Representations.
  26. Regularization and variable selection via the elastic net. Journal of the Royal Statistical Society Series B: Statistical Methodology, 67(2): 301–320.

Summary

We haven't generated a summary for this paper yet.