Accelerate Presolve in Large-Scale Linear Programming via Reinforcement Learning
Abstract: Large-scale LP problems from industry usually contain much redundancy that severely hurts the efficiency and reliability of solving LPs, making presolve (i.e., the problem simplification module) one of the most critical components in modern LP solvers. However, how to design high-quality presolve routines -- that is, the program determining (P1) which presolvers to select, (P2) in what order to execute, and (P3) when to stop -- remains a highly challenging task due to the extensive requirements on expert knowledge and the large search space. Due to the sequential decision property of the task and the lack of expert demonstrations, we propose a simple and efficient reinforcement learning (RL) framework -- namely, reinforcement learning for presolve (RL4Presolve) -- to tackle (P1)-(P3) simultaneously. Specifically, we formulate the routine design task as a Markov decision process and propose an RL framework with adaptive action sequences to generate high-quality presolve routines efficiently. Note that adaptive action sequences help learn complex behaviors efficiently and adapt to various benchmarks. Experiments on two solvers (open-source and commercial) and eight benchmarks (real-world and synthetic) demonstrate that RL4Presolve significantly and consistently improves the efficiency of solving large-scale LPs, especially on benchmarks from industry. Furthermore, we optimize the hard-coded presolve routines in LP solvers by extracting rules from learned policies for simple and efficient deployment to Huawei's supply chain. The results show encouraging economic and academic potential for incorporating machine learning to modern solvers.
- A. Brearley, G. Mitra, and H. P. Williams, “Analysis of mathematical programming problems prior to applying the simplex algorithm,” Mathematical programming, vol. 8, no. 1, pp. 54–83, 1975.
- D. L. Applegate, M. Díaz, O. Hinder, H. Lu, M. Lubin, B. O’Donoghue, and W. Schudy, “Practical large-scale linear programming using primal-dual hybrid gradient,” in Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, NeurIPS 2021, December 6-14, 2021, virtual, M. Ranzato, A. Beygelzimer, Y. N. Dauphin, P. Liang, and J. W. Vaughan, Eds., 2021, pp. 20 243–20 257.
- D. Klingman, A. Napier, and J. Stutz, “Netgen: A program for generating large scale capacitated assignment, transportation, and minimum cost flow network problems,” Management Science, vol. 20, no. 5, pp. 814–821, 1974.
- X. Li, Q. Qu, F. Zhu, J. Zeng, M. Yuan, K. Mao, and J. Wang, “Learning to reformulate for linear programming,” CoRR, vol. abs/2201.06216, 2022.
- J. Forrest, S. Vigerske, T. Ralphs, L. Hafer, J. Forrest, jpfasano, H. G. Santos, M. Saltzman, Jan-Willem, B. Kristjansson, h-i gassmann, A. King, pobonomo, S. Brito, and to st, “coin-or/clp: Release releases/1.17.7,” Jan. 2022.
- Gurobi, “Gurobi Optimizer Reference Manual,” 2022.
- Huawei Cloud, “Optverse solver, https://www.huaweicloud.com/product/modelarts/optverse.html,” 2022.
- E. D. Andersen and K. D. Andersen, “Presolving in linear programming,” Mathematical Programming, vol. 71, no. 2, pp. 221–245, 1995.
- C. Mészáros and U. H. Suhl, “Advanced preprocessing techniques for linear and quadratic programming,” OR Spectr., vol. 25, no. 4, pp. 575–595, 2003.
- Y. Bengio, A. Lodi, and A. Prouvost, “Machine learning for combinatorial optimization: A methodological tour d’horizon,” Eur. J. Oper. Res., vol. 290, no. 2, pp. 405–421, 2021.
- M. Gasse, D. Chételat, N. Ferroni, L. Charlin, and A. Lodi, “Exact combinatorial optimization with graph convolutional neural networks,” in Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, December 8-14, 2019, Vancouver, BC, Canada, H. M. Wallach, H. Larochelle, A. Beygelzimer, F. d’Alché-Buc, E. B. Fox, and R. Garnett, Eds., 2019, pp. 15 554–15 566.
- H. He, H. D. III, and J. Eisner, “Learning to search in branch and bound algorithms,” in Advances in Neural Information Processing Systems 27: Annual Conference on Neural Information Processing Systems 2014, December 8-13 2014, Montreal, Quebec, Canada, Z. Ghahramani, M. Welling, C. Cortes, N. D. Lawrence, and K. Q. Weinberger, Eds., 2014, pp. 3293–3301.
- Y. Tang, S. Agrawal, and Y. Faenza, “Reinforcement learning for integer programming: Learning to cut,” in Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13-18 July 2020, Virtual Event, ser. Proceedings of Machine Learning Research, vol. 119. PMLR, 2020, pp. 9367–9376.
- J. Zhang, C. Liu, X. Li, H. Zhen, M. Yuan, Y. Li, and J. Yan, “A survey for solving mixed integer programming via machine learning,” Neurocomputing, vol. 519, pp. 205–217, 2023.
- T. Chen, X. Chen, W. Chen, H. Heaton, J. Liu, Z. Wang, and W. Yin, “Learning to optimize: A primer and a benchmark,” Journal of Machine Learning Research, vol. 23, no. 189, pp. 1–59, 2022.
- C. Berner, G. Brockman, B. Chan, V. Cheung, P. Debiak, C. Dennison, D. Farhi, Q. Fischer, S. Hashme, C. Hesse, R. Józefowicz, S. Gray, C. Olsson, J. Pachocki, M. Petrov, H. P. de Oliveira Pinto, J. Raiman, T. Salimans, J. Schlatter, J. Schneider, S. Sidor, I. Sutskever, J. Tang, F. Wolski, and S. Zhang, “Dota 2 with large scale deep reinforcement learning,” CoRR, vol. abs/1912.06680, 2019.
- I. Sutskever, O. Vinyals, and Q. V. Le, “Sequence to sequence learning with neural networks,” in Advances in Neural Information Processing Systems 27: Annual Conference on Neural Information Processing Systems 2014, December 8-13 2014, Montreal, Quebec, Canada, Z. Ghahramani, M. Welling, C. Cortes, N. D. Lawrence, and K. Q. Weinberger, Eds., 2014, pp. 3104–3112.
- J. Schulman, F. Wolski, P. Dhariwal, A. Radford, and O. Klimov, “Proximal policy optimization algorithms,” arXiv preprint arXiv:1707.06347, 2017.
- E. B. Khalil, P. L. Bodic, L. Song, G. L. Nemhauser, and B. Dilkina, “Learning to branch in mixed integer programming,” in Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, February 12-17, 2016, Phoenix, Arizona, USA, D. Schuurmans and M. P. Wellman, Eds. AAAI Press, 2016, pp. 724–731.
- P. Gupta, E. B. Khalil, D. Chételat, M. Gasse, Y. Bengio, A. Lodi, and M. P. Kumar, “Lookback for learning to branch,” CoRR, vol. abs/2206.14987, 2022.
- A. Chmiela, E. Khalil, A. Gleixner, A. Lodi, and S. Pokutta, “Learning to schedule heuristics in branch and bound,” in Advances in Neural Information Processing Systems, M. Ranzato, A. Beygelzimer, Y. Dauphin, P. Liang, and J. W. Vaughan, Eds., vol. 34. Curran Associates, Inc., 2021, pp. 24 235–24 246.
- Z. Wang, X. Li, J. Wang, Y. Kuang, M. Yuan, J. Zeng, Y. Zhang, and F. Wu, “Learning cut selection for mixed-integer linear programming via hierarchical sequence model,” in The Eleventh International Conference on Learning Representations, 2023. [Online]. Available: https://openreview.net/forum?id=Zob4P9bRNcK
- C. Chi, A. M. Aboussalah, E. B. Khalil, J. Wang, and Z. Sherkat-Masoumi, “A deep reinforcement learning framework for column generation,” CoRR, vol. abs/2206.02568, 2022.
- O. Vinyals, M. Fortunato, and N. Jaitly, “Pointer networks,” in Advances in Neural Information Processing Systems 28: Annual Conference on Neural Information Processing Systems 2015, December 7-12, 2015, Montreal, Quebec, Canada, C. Cortes, N. D. Lawrence, D. D. Lee, M. Sugiyama, and R. Garnett, Eds., 2015, pp. 2692–2700.
- N. I. M. Gould and P. L. Toint, “Preprocessing for quadratic programming,” Math. Program., vol. 100, no. 1, pp. 95–132, 2004.
- Q. Zhou, Y. Kuang, Z. Qiu, H. Li, and J. Wang, “Promoting stochasticity for expressive policies via a simple and efficient regularization method,” in Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual, H. Larochelle, M. Ranzato, R. Hadsell, M. Balcan, and H. Lin, Eds., 2020.
- L. K. Ivert and P. Jonsson, “The potential benefits of advanced planning and scheduling systems in sales and operations planning,” Industrial Management & Data Systems, 2010.
- Z. Chen, J. Liu, X. Wang, J. Lu, and W. Yin, “On representing linear programs by graph neural networks,” CoRR, vol. abs/2209.12288, 2022.
- J. You, Z. Ying, and J. Leskovec, “Design space for graph neural networks,” in Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual, H. Larochelle, M. Ranzato, R. Hadsell, M. Balcan, and H. Lin, Eds., 2020.
- L. Engstrom, A. Ilyas, S. Santurkar, D. Tsipras, F. Janoos, L. Rudolph, and A. Madry, “Implementation matters in deep rl: A case study on ppo and trpo,” in International conference on learning representations, 2019.
- X. Li, X. Han, Z. Zhou, M. Yuan, J. Zeng, and J. Wang, “Grassland: A rapid algebraic modeling system for million-variable optimization,” in Proceedings of the 30th ACM International Conference on Information & Knowledge Management, 2021, pp. 3925–3934.
- X. Li, Q. Qu, F. Zhu, J. Zeng, M. Yuan, K. Mao, and J. Wang, “Learning to reformulate for linear programming,” arXiv preprint arXiv:2201.06216, 2022.
- D. J. Papageorgiou, G. L. Nemhauser, J. Sokol, M.-S. Cheon, and A. B. Keha, “Mirplib – a library of maritime inventory routing problem instances: Survey, core model, and benchmark results,” European Journal of Operational Research, vol. 235, no. 2, pp. 350–366, 2014, maritime Logistics.
- G. Cornuejols, R. Sridharan, and J. Thizy, “A comparison of heuristics and relaxations for the capacitated plant location problem,” European Journal of Operational Research, vol. 50, no. 3, pp. 280–297, 1991.
- M. Hewitt, G. L. Nemhauser, and M. W. P. Savelsbergh, “Combining exact and heuristic approaches for the capacitated fixed-charge network flow problem,” INFORMS J. Comput., vol. 22, no. 2, pp. 314–325, 2010.
- A. G. Labassi, D. Chételat, and A. Lodi, “Learning to compare nodes in branch and bound with graph neural networks,” in Advances in Neural Information Processing Systems.
- T. Achterberg, “Constraint integer programming,” Ph.D. dissertation, Berlin Institute of Technology, 2007.
- P. Gemander, W. Chen, D. Weninger, L. Gottwald, A. M. Gleixner, and A. Martin, “Two-row and two-column mixed-integer presolve using hashing-based pairing methods,” EURO J. Comput. Optim., vol. 8, no. 3, pp. 205–240, 2020.
- R. Yang, J. Wang, Z. Geng, M. Ye, S. Ji, B. Li, and F. Wu, “Learning task-relevant representations for generalization via characteristic functions of reward sequence distributions,” in KDD ’22: The 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Washington, DC, USA, August 14 - 18, 2022, A. Zhang and H. Rangwala, Eds. ACM, 2022, pp. 2242–2252.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.