Papers
Topics
Authors
Recent
Search
2000 character limit reached

Robust Losses for Decision-Focused Learning

Published 6 Oct 2023 in cs.LG and math.OC | (2310.04328v2)

Abstract: Optimization models used to make discrete decisions often contain uncertain parameters that are context-dependent and estimated through prediction. To account for the quality of the decision made based on the prediction, decision-focused learning (end-to-end predict-then-optimize) aims at training the predictive model to minimize regret, i.e., the loss incurred by making a suboptimal decision. Despite the challenge of the gradient of this loss w.r.t. the predictive model parameters being zero almost everywhere for optimization problems with a linear objective, effective gradient-based learning approaches have been proposed to minimize the expected loss, using the empirical loss as a surrogate. However, empirical regret can be an ineffective surrogate because empirical optimal decisions can vary substantially from expected optimal decisions. To understand the impact of this deficiency, we evaluate the effect of aleatoric and epistemic uncertainty on the accuracy of empirical regret as a surrogate. Next, we propose three novel loss functions that approximate expected regret more robustly. Experimental results show that training two state-of-the-art decision-focused learning approaches using robust regret losses improves test-sample empirical regret in general while keeping computational time equivalent relative to the number of training epochs.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (25)
  1. Yoshua Bengio. Using a financial training criterion rather than a prediction criterion. International Journal of Neural Systems, 8:433–443, 1997. doi:10.1142/s0129065797000422.
  2. Smart "predict, then optimize". Management Science, 68:9–26, 2022. ISSN 15265501. doi:10.1287/mnsc.2020.3922.
  3. Differentiation of blackbox combinatorial solvers. International Conference on Learning Representations, 2019. URL http://arxiv.org/abs/1912.02175.
  4. Learning with differentiable perturbed optimizers. Advances in Neural Information Processing Systems, 33:9508–9519, 2020.
  5. Aleatory or epistemic? Does it matter? Structural Safety, 31:105–112, 2009. ISSN 01674730. doi:10.1016/j.strusafe.2008.06.020.
  6. Contrastive losses and solution caching for predict-and-optimize. arXiv preprint arXiv:2011.05354, 2020. URL http://arxiv.org/abs/2011.05354.
  7. Theory and applications of robust optimization. SIAM Review, 53:464–501, 2011. ISSN 00361445. doi:10.1137/080734510.
  8. Robust and adaptive optimization. Dynamic Ideas LLC, 2022. ISBN 9781733788526.
  9. Philip E Cheng. Strong consistency of nearest neighbor regression function estimators. Journal of Multivariate Analysis, 15:63–72, 1984. doi:10.1016/0047-259X(84)90067-8.
  10. Heinrich Jiang. Non-asymptotic uniform rates of consistency for k-nn regression. Proceedings of the AAAI Conference on Artificial Intelligence, 33:3999–4006, 2019. doi:10.1609/aaai.v33i01.33013999.
  11. Neural networks and the bias/variance dilemma. Neural computation, 4:1–58, 1992. doi:10.1162/neco.1992.4.1.1.
  12. Bo Tang and Elias B. Khalil. Pyepo: A pytorch-based end-to-end predict-then-optimize library for linear and integer programming. arXiv preprint arXiv:2206.14234, 2022. URL http://arxiv.org/abs/2206.14234.
  13. Gurobi Optimization, LLC. Gurobi optimizer reference manual, 2021. URL https://www.gurobi.com. Accessed: 2023-08-15.
  14. Csplib problem 059: Energy-cost aware scheduling, 1999. URL http://www.csplib.org/Problems/prob059. Accessed: 2023-08-15.
  15. Decision-focused learning: Through the lens of learning to rank. International Conference on Machine Learning, pages 14935–14947, 2022. URL https://proceedings.mlr.press/v162/mandi22a.html.
  16. Properties of energy-price forecasts for scheduling. International Conference on Principles and Practice of Constraint Programming, pages 957–972, 2012. doi:10.1007/978-3-642-33558-7_68.
  17. Brandon Amos and JZ Kolter. Optnet: Differentiable optimization as a layer in neural networks. International Conference on Machine Learning, pages 136–145, 2017. URL https://proceedings.mlr.press/v70/amos17a.html.
  18. Melding the data-decisions pipeline: Decision-focused learning for combinatorial optimization. Proceedings of the AAAI Conference on Artificial Intelligence, 33:1658–1665, 2019. doi:10.1609/aaai.v33i01.33011658.
  19. Interior point solving for lp-based prediction+optimisation. Advances in Neural Information Processing Systems, 33:7272–7282, 2020. URL http://arxiv.org/abs/2010.13943.
  20. Mipaal: Mixed integer program as a layer. Proceedings of the AAAI Conference on Artificial Intelligence, 34:1504–1511, 2020. doi:10.1609/aaai.v34i02.5509.
  21. Score function gradient estimation to widen the applicability of decision-focused learning. arXiv preprint arXiv:2307.05213, 2023. URL http://arxiv.org/abs/2307.05213.
  22. Decision-focused learning without differentiable optimization: Learning locally optimized decision losses. arXiv preprint arXiv:2203.16067, 2022. URL http://arxiv.org/abs/2203.16067.
  23. A survey of contextual optimization methods for decision making under uncertainty. arXiv preprint arXiv:2306.10374, 2023. URL http://arxiv.org/abs/2306.10374.
  24. Predict-then-calibrate: A new perspective of robust contextual lp. arXiv preprint arXiv:2305.15686, 2023. URL http://arxiv.org/abs/2305.15686.
  25. Regularization for deep learning: A taxonomy. arXiv preprint arXiv:1710.10686, 2017. URL http://arxiv.org/abs/1710.10686.
Citations (2)

Summary

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.