Epsilon-Greedy Thompson Sampling to Bayesian Optimization (2403.00540v3)
Abstract: Bayesian optimization (BO) has become a powerful tool for solving simulation-based engineering optimization problems thanks to its ability to integrate physical and mathematical understandings, consider uncertainty, and address the exploitation-exploration dilemma. Thompson sampling (TS) is a preferred solution for BO to handle the exploitation-exploration trade-off. While it prioritizes exploration by generating and minimizing random sample paths from probabilistic models -- a fundamental ingredient of BO -- TS weakly manages exploitation by gathering information about the true objective function after it obtains new observations. In this work, we improve the exploitation of TS by incorporating the $\varepsilon$-greedy policy, a well-established selection strategy in reinforcement learning. We first delineate two extremes of TS, namely the generic TS and the sample-average TS. The former promotes exploration, while the latter favors exploitation. We then adopt the $\varepsilon$-greedy policy to randomly switch between these two extremes. Small and large values of $\varepsilon$ govern exploitation and exploration, respectively. By minimizing two benchmark functions and solving an inverse problem of a steel cantilever beam, we empirically show that $\varepsilon$-greedy TS equipped with an appropriate $\varepsilon$ is more robust than its two extremes, matching or outperforming the better of the generic TS and the sample-average TS.
- Practical Bayesian optimization of machine learning algorithms. In Advances in Neural Information Processing Systems, volume 25, pages 2951–2959. Curran Associates, Inc., 2012. URL https://proceedings.neurips.cc/paper_files/paper/2012/file/05311655a15b75fab86956663e1819cd-Paper.pdf. Cited on page 2.
- Peter I. Frazier. Bayesian optimization. INFORMS TutORials in Operations Research, pages 255–278. INFORMS, October 2018. doi:10.1287/educ.2018.0188. URL https://doi.org/10.1287/educ.2018.0188. Cited on page 255.
- Roman Garnett. Bayesian optimization. Cambridge University Press, Cambridge, 2023. ISBN 9781108425780. Cited on page 1.
- Multi-fidelity bayesian optimization in engineering design. arXiv, November 2023. URL https://arxiv.org/abs/2311.13050.
- Probabilistic numerics: Computation as machine learning. Cambridge University Press, 2022. ISBN 1316730336. Cited on chapter 32.
- An informational approach to the global optimization of expensive-to-evaluate functions. Journal of Global Optimization, 44(4):509–534, 2009. ISSN 1573-2916. doi:10.1007/s10898-008-9354-2. URL https://doi.org/10.1007/s10898-008-9354-2. Cited on pages 8, 9.
- Entropy search for information-efficient global optimization. Journal of Machine Learning Research, 13(6):1809–1837, 2012. ISSN 1532-4435. URL https://www.jmlr.org/papers/v13/hennig12a.html. Cited on page 2.
- Predictive entropy search for efficient global optimization of black-box functions. In Advances in Neural Information Processing Systems, volume 27, pages 918–926. Curran Associates, Inc., 2014. URL https://proceedings.neurips.cc/paper_files/paper/2014/hash/069d3bb002acd8d7dd095917f9efe4cb-Abstract.html. Cited on page 3.
- Zi Wang and Stefanie Jegelka. Max-value entropy search for efficient Bayesian optimization. In Proceedings of the 34th International Conference on Machine Learning, volume 70, pages 3627–3635. PMLR, 2017. URL https://proceedings.mlr.press/v70/wang17e.html. Cited on pages 2, 3, 4.
- Efficient global optimization of expensive black-box functions. Journal of Global Optimization, 13(4):455–492, 1998. ISSN 1573-2916. doi:10.1023/A:1008306431147. URL https://doi.org/10.1023/A:1008306431147. Cited on page 17.
- On the design of optimization strategies based on global response surface approximation models. Journal of Global Optimization, 33(1):31–59, 2005. ISSN 1573-2916. doi:10.1007/s10898-004-6733-1. URL https://doi.org/10.1007/s10898-004-6733-1. Cited on page 9.
- Gaussian process optimization in the bandit setting: No regret and experimental design. In Proceedings of the 27th International Conference on International Conference on Machine Learning, pages 1015–1022. Omnipress, 2010. URL https://icml.cc/Conferences/2010/papers/422.pdf. Cited on page 4.
- A knowledge-gradient policy for sequential information collection. SIAM Journal on Control and Optimization, 47(5):2410–2439, January 2008. ISSN 0363-0129. doi:10.1137/070693424. URL https://doi.org/10.1137/070693424. Cited on page 2.
- An empirical evaluation of Thompson sampling. In Advances in Neural Information Processing Systems, volume 24, pages 2249–2257. Curran Associates, Inc., 2011. URL https://papers.nips.cc/paper_files/paper/2011/hash/e53a0a2978c28872a4505bdb51db06dc-Abstract.html. Cited on page 2.
- A tutorial on Thompson sampling. Foundations and Trends® in Machine Learning, 11(1):1–96, 2018. ISSN 1935-8237. doi:10.1561/2200000070. URL http://dx.doi.org/10.1561/2200000070. Cited on page 7.
- Parallelised bayesian optimisation via thompson sampling. In Proceedings of the Twenty-First International Conference on Artificial Intelligence and Statistics, volume 84 of Proceedings of Machine Learning Research, pages 133–142. PMLR, 09–11 Apr 2018. URL https://proceedings.mlr.press/v84/kandasamy18a.html. Cited on page 4.
- Reinforcement learning: An introduction. MIT Press, 2018. ISBN 9780262039246. Cited on pages 28, 30, 32.
- Greed Is Good: Exploration and Exploitation Trade-offs in Bayesian Optimisation. ACM Transactions on Evolutionary Learning and Optimization, 1(1), apr 2021. ISSN 2688-3007. doi:10.1145/3425501. URL https://doi.org/10.1145/3425501. Cited on pages 10, 11, 21.
- Thompson sampling with less exploration is fast and optimal. In Proceedings of the 40th International Conference on Machine Learning, volume 202 of Proceedings of Machine Learning Research, pages 15239–15261. PMLR, 23–29 Jul 2023. URL https://proceedings.mlr.press/v202/jin23b.html. Cited on pages 3, 4.
- Carl Edward Rasmussen and Christopher K I Williams. Gaussian processes for machine learning. The MIT Press, Massachusetts, USA, 2006. ISBN 9780521872508. doi:10.7551/mitpress/3206.001.0001. URL https://doi.org/10.7551/mitpress/3206.001.0001. Cited on page 2.
- C. Bishop. Pattern recognition and machine learning. Springer, Berkeley, CA, USA, 2006. ISBN 9780387310732. Cited on page 87.
- Holger Wendland. Scattered data approximation, volume 17. Cambridge University Press, 2004. ISBN 9780511264320. Cited on page 70.
- Practical Hilbert space approximate Bayesian Gaussian processes for probabilistic programming. Statistics and Computing, 33(1):17, 2022. ISSN 1573-1375. doi:10.1007/s11222-022-10167-2. URL https://doi.org/10.1007/s11222-022-10167-2. Cited on page 4.
- Random features for large-scale kernel machines. In Advances in Neural Information Processing Systems, volume 20. Curran Associates, Inc., 2007. URL https://proceedings.neurips.cc/paper_files/paper/2007/file/013a006f03dbc5392effeb8f18fda755-Paper.pdf. Cited on page 1.
- Efficiently sampling functions from Gaussian process posteriors. In Proceedings of the 37th International Conference on Machine Learning, volume 119, pages 10292–10302. PMLR, 2020. URL https://proceedings.mlr.press/v119/wilson20a.html. Cited on pages 2, 3.
- BoTorch: A Framework for Efficient Monte-Carlo Bayesian Optimization. In Advances in Neural Information Processing Systems, volume 33, pages 21524–21538. Curran Associates, Inc., 2020. URL https://proceedings.neurips.cc/paper_files/paper/2020/file/f5b1b89d98b7286673128a5fb112cb9a-Paper.pdf. Cited on pages 2, 3.
- Efficient multiobjective optimization employing Gaussian processes, spectral sampling and a genetic algorithm. Journal of Global Optimization, 71(2):407–438, 2018. ISSN 1573-2916. doi:10.1007/s10898-018-0609-2. URL https://doi.org/10.1007/s10898-018-0609-2. Cited on page 14.
- Additive scaling and the DIRECT algorithm. Journal of Global Optimization, 36(4):597–608, 2006. ISSN 1573-2916. doi:10.1007/s10898-006-9029-9. URL https://doi.org/10.1007/s10898-006-9029-9. Cited on pages 2, 3.
- Engineering design via surrogate modelling: A practical guide. John Wiley & Sons, West Sussex, UK, 2008. ISBN 0470770791. Cited on page 15.
- A framework for Bayesian optimization in embedded subspaces. In Proceedings of the 36th International Conference on Machine Learning, volume 97, pages 4752–4761. PMLR, 09–15 Jun 2019. URL https://proceedings.mlr.press/v97/nayebi19a.html. Cited on pages 2, 3.
- Scalable global optimization via local Bayesian optimization. In Advances in Neural Information Processing Systems, volume 32, pages 5496–5507. Curran Associates, Inc., 2019. URL https://proceedings.neurips.cc/paper_files/paper/2019/hash/6c990b7aca7bc7058f5e98ea909e924b-Abstract.html. Cited on pages 3, 4.
- S Surjanovic and D Bingham. Virtual library of simulation experiments: Test functions and datasets, 2013. URL http://www.sfu.ca/~ssurjano.