Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
169 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Locally Convex Global Loss Network for Decision-Focused Learning (2403.01875v2)

Published 4 Mar 2024 in cs.LG and cs.AI

Abstract: In decision-making problem under uncertainty, predicting unknown parameters is often considered independent of the optimization part. Decision-focused Learning (DFL) is a task-oriented framework to integrate prediction and optimization by adapting predictive model to give better decision for the corresponding task. Here, an inevitable challenge arises when computing gradients of the optimal decision with respect to the parameters. Existing researches cope this issue by smoothly reforming surrogate optimization or construct surrogate loss function that mimic task loss. However, they are applied to restricted optimization domain. In this paper, we propose Locally Convex Global Loss Network (LCGLN), a global surrogate loss model which can be implemented in a general DFL paradigm. LCGLN learns task loss via partial input convex neural network which is guaranteed to be convex for chosen inputs, while keeping the non-convex global structure for the other inputs. This enables LCGLN to admit general DFL through only a single surrogate loss without any sense for choosing appropriate parametric forms. We confirm effectiveness and flexibility of LCGLN by evaluating our proposed model with three stochastic decision-making problems.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (26)
  1. Differentiable convex optimization layers. Advances in neural information processing systems, 32, 2019.
  2. Optnet: Differentiable optimization as a layer in neural networks. In International Conference on Machine Learning, pp.  136–145. PMLR, 2017.
  3. Input convex neural networks. In International Conference on Machine Learning, pp.  146–155. PMLR, 2017.
  4. The limited multi-label projection layer. arXiv preprint arXiv:1906.08707, 2019.
  5. Learning with differentiable pertubed optimizers. Advances in neural information processing systems, 33:9508–9519, 2020.
  6. The russell-yasuda kasai model: An asset/liability model for a japanese insurance company using multistage stochastic programming. Interfaces, 24(1):29–49, 1994.
  7. Distributionally robust optimization under moment uncertainty with application to data-driven problems. Operations research, 58(3):595–612, 2010.
  8. Task-based end-to-end model learning in stochastic optimization. Advances in neural information processing systems, 30, 2017.
  9. Smart “predict, then optimize”. Management Science, 68(1):9–26, 2022.
  10. Mipaal: Mixed integer program as a layer. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pp.  1504–1511, 2020.
  11. Short-term hydropower production planning by stochastic programming. Computers & Operations Research, 35(8):2656–2671, 2008.
  12. Portfolio selection with parameter and model uncertainty: A multi-prior approach. The Review of Financial Studies, 20(1):41–81, 2007.
  13. Noise-contrastive estimation: A new estimation principle for unnormalized statistical models. In Proceedings of the thirteenth international conference on artificial intelligence and statistics, pp.  297–304. JMLR Workshop and Conference Proceedings, 2010.
  14. End-to-end stochastic optimization with energy-based model. Advances in Neural Information Processing Systems, 35:11341–11354, 2022.
  15. Interior point solving for lp-based prediction+ optimisation, 2020. URL http://arxiv. org/abs, 2010.
  16. Smart predict-and-optimize for hard combinatorial optimization problems. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pp.  1603–1610, 2020.
  17. Decision-focused learning: through the lens of learning to rank. In International Conference on Machine Learning, pp.  14935–14947. PMLR, 2022.
  18. Decision-focused learning: Foundations, state of the art, benchmark and future opportunities. arXiv preprint arXiv:2307.13565, 2023.
  19. Learning what’s easy: Fully differentiable neural easy-first taggers. In Proceedings of the 2017 conference on empirical methods in natural language processing, pp.  349–362, 2017.
  20. Contrastive losses and solution caching for predict-and-optimize. arXiv preprint arXiv:2011.05354, 2020.
  21. Implicit mle: backpropagating through discrete exponential family distributions. Advances in Neural Information Processing Systems, 34:14567–14579, 2021.
  22. Perturb-and-map random fields: Using discrete optimization to learn and sample from energy models. In 2011 International Conference on Computer Vision, pp.  193–200. IEEE, 2011.
  23. Decision-focused learning without decision-making: Learning locally optimized decision losses. Advances in Neural Information Processing Systems, 35:1320–1332, 2022.
  24. Multistage stochastic programming model for electric power capacity expansion problem. Japan journal of industrial and applied mathematics, 20:379–397, 2003.
  25. Melding the data-decisions pipeline: Decision-focused learning for combinatorial optimization. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, pp.  1658–1665, 2019.
  26. Differentiable top-k with optimal transport. Advances in Neural Information Processing Systems, 33:20520–20531, 2020.

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com