Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
162 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Sample Complexity of Algorithm Selection Using Neural Networks and Its Applications to Branch-and-Cut (2402.02328v3)

Published 4 Feb 2024 in cs.LG and math.OC

Abstract: Data-driven algorithm design is a paradigm that uses statistical and machine learning techniques to select from a class of algorithms for a computational problem an algorithm that has the best expected performance with respect to some (unknown) distribution on the instances of the problem. We build upon recent work in this line of research by considering the setup where, instead of selecting a single algorithm that has the best performance, we allow the possibility of selecting an algorithm based on the instance to be solved, using neural networks. In particular, given a representative sample of instances, we learn a neural network that maps an instance of the problem to the most appropriate algorithm for that instance. We formalize this idea and derive rigorous sample complexity bounds for this learning problem, in the spirit of recent work in data-driven algorithm design. We then apply this approach to the problem of making good decisions in the branch-and-cut framework for mixed-integer optimization (e.g., which cut to add?). In other words, the neural network will take as input a mixed-integer optimization instance and output a decision that will result in a small branch-and-cut tree for that instance. Our computational results provide evidence that our particular way of using neural networks for cut selection can make a significant impact in reducing branch-and-cut tree sizes, compared to previous data-driven approaches.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (27)
  1. Maria-Florina Balcan. Data-driven algirithm design. In Tim Roughgarden, editor, Beyond the Worst Case Analysis of Algorithms. Cambridge University Press, 2020.
  2. How much data is sufficient to learn high-performing algorithms? generalization guarantees for data-driven algorithm design. In Proceedings of the 53rd Annual ACM SIGACT Symposium on Theory of Computing, pages 919–932, 2021a.
  3. Algorithms with predictions. Communications of the ACM, 65(7):33–35, 2022.
  4. Alexander Schrijver. Theory of Linear and Integer Programming. John Wiley and Sons, New York, 1986.
  5. Integer and combinatorial optimization, volume 18. Wiley New York, 1988.
  6. Integer programming, volume 271. Springer, 2014.
  7. Sample complexity of tree search configuration: Cutting planes and beyond. Advances in Neural Information Processing Systems, 34:4015–4027, 2021b.
  8. Improved sample complexity bounds for branch-and-cut. arXiv preprint arXiv:2111.11207, 2021c.
  9. Structural analysis of branch-and-cut and the learnability of gomory mixed integer cuts. Advances in Neural Information Processing Systems, 35:33890–33903, 2022.
  10. Learning to branch. In International conference on machine learning, pages 344–353. PMLR, 2018.
  11. Machine learning for cutting planes in integer programming: A survey. arXiv preprint arXiv:2302.09166, 2023.
  12. Reinforcement learning for integer programming: Learning to cut. In International conference on machine learning, pages 9367–9376. PMLR, 2020.
  13. Learning to select cuts for efficient mixed-integer programming. Pattern Recognition, 123:108353, 2022.
  14. Adaptive cut selection in mixed-integer linear programming. Open Journal of Mathematical Optimization, 4:1–28, 2023.
  15. Neural network learning: Theoretical foundations, volume 9. cambridge university press Cambridge, 1999.
  16. Neural networks with linear threshold activations: structure and algorithms. Mathematical Programming, pages 1–24, 2023.
  17. Multidimensional knapsack problems. Springer, 2004.
  18. Vasek Chvátal. Hard knapsack problems. Operations Research, 28(6):1402–1411, 1980.
  19. Deterministic policy gradient algorithms. In International conference on machine learning, pages 387–395. Pmlr, 2014.
  20. Addressing function approximation error in actor-critic methods. In International conference on machine learning, pages 1587–1596. PMLR, 2018.
  21. Gurobi Optimization, LLC. Gurobi Optimizer Reference Manual, 2023. URL https://www.gurobi.com.
  22. The scip optimization suite 7.0. 2020.
  23. Mixed 0-1 programming by lift-and-project in a branch-and-cut framework. Management Science, 42(9):1229–1246, 1996.
  24. Tobias Achterberg. Constraint integer programming. 2007.
  25. Estimating or propagating gradients through stochastic neurons for conditional computation. arXiv preprint arXiv:1308.3432, 2013.
  26. Jiri Matousek. Geometric discrepancy: An illustrated guide, volume 18. Springer Science & Business Media, 1999.
  27. Nearly-tight vc-dimension and pseudodimension bounds for piecewise linear neural networks. The Journal of Machine Learning Research, 20(1):2285–2301, 2019.

Summary

We haven't generated a summary for this paper yet.