Multi-Fidelity Bayesian Optimization With Across-Task Transferable Max-Value Entropy Search (2403.09570v4)
Abstract: In many applications, ranging from logistics to engineering, a designer is faced with a sequence of optimization tasks for which the objectives are in the form of black-box functions that are costly to evaluate. Furthermore, higher-fidelity evaluations of the optimization objectives often entail a larger cost. Existing multi-fidelity black-box optimization strategies select candidate solutions and fidelity levels with the goal of maximizing the information about the optimal value or the optimal solution for the current task. Assuming that successive optimization tasks are related, this paper introduces a novel information-theoretic acquisition function that balances the need to acquire information about the current task with the goal of collecting information transferable to future tasks. The proposed method transfers across tasks distributions over parameters of a Gaussian process surrogate model by implementing particle-based variational Bayesian updates. Theoretical insights based on the analysis of the expected regret substantiate the benefits of acquiring transferable knowledge across tasks. Furthermore, experimental results across synthetic and real-world examples reveal that the proposed acquisition strategy that caters to future tasks can significantly improve the optimization efficiency as soon as a sufficient number of tasks is processed.
- M. Zhang, H. Li, S. Pan, J. Lyu, S. Ling, and S. Su, “Convolutional neural networks-based lung nodule classification: A surrogate-assisted evolutionary algorithm for hyperparameter optimization,” IEEE Transactions on Evolutionary Computation, vol. 25, no. 5, pp. 869–882, 2021.
- L. Demetrio, B. Biggio, G. Lagorio, F. Roli, and A. Armando, “Functionality-preserving black-box optimization of adversarial Windows malware,” IEEE Transactions on Information Forensics and Security, vol. 16, pp. 3469–3478, 2021.
- J. Zhou, Z. Yang, Y. Si, L. Kang, H. Li, M. Wang, and Z. Zhang, “A trust-region parallel Bayesian optimization method for simulation-driven antenna design,” IEEE Transactions on Antennas and Propagation, vol. 69, no. 7, pp. 3966–3981, 2020.
- H. B. Moss, V. Aggarwal, N. Prateek, J. González, and R. Barra-Chicote, “Boffin tts: Few-shot speaker adaptation by Bayesian optimization,” in Proceedings of IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2020.
- S. Xu, J. Li, P. Cai, X. Liu, B. Liu, and X. Wang, “Self-improving photosensitizer discovery system via Bayesian search with first-principle simulations,” Journal of the American Chemical Society, vol. 143, no. 47, pp. 19 769–19 777, 2021.
- L. Maggi, A. Valcarce, and J. Hoydis, “Bayesian optimization for radio resource management: Open loop power control,” IEEE Journal on Selected Areas in Communications, vol. 39, no. 7, pp. 1858–1871, 2021.
- Y. Zhang, O. Simeone, S. T. Jose, L. Maggi, and A. Valcarce, “Bayesian and multi-armed contextual meta-optimization for efficient wireless radio resource management,” IEEE Transactions on Cognitive Communications and Networking, vol. 9, no. 5, pp. 1282–1295, 2023.
- Y. Wang, J. Fang, Y. Cheng, H. She, Y. Guo, and G. Zheng, “Cooperative end-edge-cloud computing and resource allocation for digital twin enabled 6G industrial iot,” IEEE Journal of Selected Topics in Signal Processing, pp. 1–14, 2023.
- J. Johnston, X.-Y. Liu, S. Wu, and X. Wang, “A curriculum learning approach to optimization with application to downlink beamforming,” IEEE Transactions on Signal Processing, vol. 72, pp. 84–98, 2024.
- R. Highfield and P. Coveney, “Virtual you: how building your digital twin will revolutionize medicine and change your life,” Princeton University Press, 2023.
- J. Hoydis, F. A. Aoudia, S. Cammerer, M. Nimier-David, N. Binder, G. Marcus, and A. Keller, “Sionna rt: Differentiable ray tracing for radio propagation modeling,” arXiv preprint arXiv:2303.11103, 2023.
- C. Ruah, O. Simeone, and B. Al-Hashimi, “A Bayesian framework for digital twin-based control, monitoring, and data collection in wireless systems,” IEEE Journal on Selected Areas in Communications, vol. 41, no. 10, pp. 3146–3160, 2023.
- C. Ruah, O. Simeone, J. Hoydis, and B. Al-Hashimi, “Calibrating wireless ray tracing for digital twinning using local phase error estimates,” arXiv preprint arXiv:2312.12625, 2023.
- W. Chen, W. Hong, H. Zhang, P. Yang, and K. Tang, “Multi-fidelity simulation modeling for discrete event simulation: An optimization perspective,” IEEE Transactions on Automation Science and Engineering, vol. 20, no. 2, pp. 1156–1169, 2022.
- D. R. Jones, M. Schonlau, and W. J. Welch, “Efficient global optimization of expensive black-box functions,” Journal of Global optimization, vol. 13, pp. 455–492, 1998.
- P. I. Frazier, W. B. Powell, and S. Dayanik, “A knowledge-gradient policy for sequential information collection,” SIAM Journal on Control and Optimization, vol. 47, no. 5, pp. 2410–2439, 2008.
- Z. Wang, V. Y. Tan, and J. Scarlett, “Tight regret bounds for noisy optimization of a Brownian motion,” IEEE Transactions on Signal Processing, vol. 70, pp. 1072–1087, 2022.
- Z. Wang and S. Jegelka, “Max-value entropy search for efficient Bayesian optimization,” in Proceedings of International Conference on Machine Learning, Sydney, Australia, 2017.
- N. Srinivas, A. Krause, S. M. Kakade, and M. W. Seeger, “Information-theoretic regret bounds for Gaussian process optimization in the bandit setting,” IEEE Transactions on Information Theory, vol. 58, no. 5, pp. 3250–3265, 2012.
- H. B. Moss, D. S. Leslie, J. Gonzalez, and P. Rayson, “Gibbon: General-purpose information-based Bayesian optimisation,” Journal of Machine Learning Research, vol. 22, no. 1, pp. 10 616–10 664, 2021.
- K. Swersky, J. Snoek, and R. P. Adams, “Multi-task Bayesian optimization,” in Proceedings of Advances in Neural Information Processing Systems, Nevada, USA, 2013.
- M. Poloczek, J. Wang, and P. Frazier, “Multi-information source optimization,” in Proceedings of Advances in neural information processing systems, California, USA, 2017.
- R. Lam, D. L. Allaire, and K. E. Willcox, “Multifidelity optimization using statistical surrogate modeling for non-hierarchical information sources,” in Proceedings of 56th AIAA/ASCE/AHS/ASC Structures, Structural Dynamics, and Materials Conference, Florida, USA, 2015.
- K. Kandasamy, G. Dasarathy, J. Schneider, and B. Póczos, “Multi-fidelity Bayesian optimisation with continuous approximations,” in Proceedings of International Conference on Machine Learning, Sydney, Australia, 2017.
- J. Wu, S. Toscano-Palmerin, P. I. Frazier, and A. G. Wilson, “Practical multi-fidelity Bayesian optimization for hyperparameter tuning,” in Uncertainty in Artificial Intelligence, 2020.
- P. Hennig and C. J. Schuler, “Entropy search for information-efficient global optimization.” Journal of Machine Learning Research, vol. 13, no. 6, 2012.
- H. B. Moss, D. S. Leslie, and P. Rayson, “Mumbo: Multi-task max-value Bayesian optimization,” in Proceedings of Joint European Conference on Machine Learning and Knowledge Discovery in Databases, Ghent, Belgium, 2020.
- S. Takeno, H. Fukuoka, Y. Tsukada, T. Koyama, M. Shiga, I. Takeuchi, and M. Karasuyama, “Multi-fidelity Bayesian optimization with max-value entropy search and its parallelization,” in Proceedings of International Conference on Machine Learning, 2020.
- P. Mikkola, J. Martinelli, L. Filstroff, and S. Kaski, “Multi-fidelity Bayesian optimization with unreliable information sources,” in Proceedings of International Conference on Artificial Intelligence and Statistics, Valencia, Spain, 2023.
- O. Vinyals, C. Blundell, T. Lillicrap, D. Wierstra et al., “Matching networks for one shot learning,” in Proceedings of Advances in Neural Information Processing Systems, vol. 29, Barcelona, Spain, 2016.
- Y. Zhang, J. Jordon, A. M. Alaa, and M. van der Schaar, “Lifelong Bayesian optimization,” arXiv preprint arXiv:1905.12280, 2019.
- J. Rothfuss, C. Koenig, A. Rupenyan, and A. Krause, “Meta-learning priors for safe Bayesian optimization,” in Proceedings of Conference on Robot Learning, Georgia, USA, 2023.
- Q. Liu and D. Wang, “Stein variational gradient descent: A general purpose Bayesian inference algorithm,” Proceedings of Advances in Neural Information Processing Systems, Barcelona, Spain, 2016.
- L. C. W. Dixon, “The global optimization problem: an introduction,” Towards Global Optimiation 2, pp. 1–15, 1978.
- H. Li and J. Zhang, “Fast source term estimation using the PGA-NM hybrid method,” Engineering Applications of Artificial Intelligence, vol. 62, pp. 68–79, 2017.
- B. Lei, T. Q. Kirk, A. Bhattacharya, D. Pati, X. Qian, R. Arroyave, and B. K. Mallick, “Bayesian optimization with adaptive surrogate models for automated experimental design,” Npj Computational Materials, vol. 7, no. 1, p. 194, 2021.
- G. Malkomes, C. Schaff, and R. Garnett, “Bayesian optimization for automated model selection,” in Proceedings of Advances in Neural Information Processing Systems, Barcelona, Spain, 2016.
- A. I. Forrester, A. Sóbester, and A. J. Keane, “Multi-fidelity optimization via surrogate modelling,” Proceedings of the royal society a: mathematical, physical and engineering sciences, vol. 463, no. 2088, pp. 3251–3269, 2007.
- L. Le Gratiet and J. Garnier, “Recursive co-kriging model for design of computer experiments with multiple levels of fidelity,” International Journal for Uncertainty Quantification, 2014.
- J. Rothfuss, V. Fortuin, M. Josifoski, and A. Krause, “PACOH: Bayes-optimal meta-learning with PAC-guarantees,” in Proceedings of International Conference on Machine Learning, 2021.
- E. V. Bonilla, K. Chai, and C. Williams, “Multi-task Gaussian process prediction,” in Proceedings of Advances in Neural Information Processing Systems, Vancouver, Canada, 2007.
- J. Gardner, G. Pleiss, K. Q. Weinberger, D. Bindel, and A. G. Wilson, “Gpytorch: Blackbox matrix-matrix Gaussian process inference with gpu acceleration,” in Proceedings of Advances in Neural Information Processing Systems, Montreal, Canada, 2018.
- M. C. Kennedy and A. O’Hagan, “Predicting the output from a complex computer code when fast approximations are available,” Biometrika, vol. 87, no. 1, pp. 1–13, 2000.
- M. A. Alvarez and N. D. Lawrence, “Computationally efficient convolved multiple output Gaussian processes,” Journal of Machine Learning Research, vol. 12, pp. 1459–1500, 2011.
- A. Damianou and N. D. Lawrence, “Deep Gaussian processes,” in Proceedings of International Conference on Artificial Intelligence and Statistics, Arizona, USA, 2013.
- S. J. Sloman, A. Bharti, and S. Kaski, “The fundamental dilemma of Bayesian active meta-learning,” arXiv preprint arXiv:2310.14968, 2023.
- K. Moshksar and A. K. Khandani, “Arbitrarily tight bounds on differential entropy of Gaussian mixtures,” IEEE Transactions on Information Theory, vol. 62, no. 6, pp. 3340–3354, 2016.
- F. Nielsen and K. Sun, “Guaranteed bounds on information-theoretic measures of univariate mixtures using piecewise log-sum-exp inequalities,” Entropy, vol. 18, no. 12, p. 442, 2016.
- S. Bubeck, R. Munos, and G. Stoltz, “Pure exploration in multi-armed bandits problems,” in Proceedings of International Conference on Algorithmic Learning Theory, Porto, Portugal, 2009.
- H. Li, Y. Jin, and T. Chai, “Evolutionary multi-objective Bayesian optimization based on multisource online transfer learning,” IEEE Transactions on Emerging Topics in Computational Intelligence, vol. 8, no. 1, pp. 488–502, 2024.
- W. J. Maddox, M. Balandat, A. G. Wilson, and E. Bakshy, “Bayesian optimization with high-dimensional outputs,” in Proceedings of Advances in Neural Information Processing Systems, 2021.
- Z. Liang, Y. Zhu, X. Wang, Z. Li, and Z. Zhu, “Evolutionary multitasking for multi-objective optimization based on generative strategies,” IEEE Transactions on Evolutionary Computation, vol. 27, no. 4, pp. 1042–1056, 2023.
Collections
Sign up for free to add this paper to one or more collections.