CMA-ES for Post Hoc Ensembling in AutoML: A Great Success and Salvageable Failure (2307.00286v1)
Abstract: Many state-of-the-art automated machine learning (AutoML) systems use greedy ensemble selection (GES) by Caruana et al. (2004) to ensemble models found during model selection post hoc. Thereby, boosting predictive performance and likely following Auto-Sklearn 1's insight that alternatives, like stacking or gradient-free numerical optimization, overfit. Overfitting in Auto-Sklearn 1 is much more likely than in other AutoML systems because it uses only low-quality validation data for post hoc ensembling. Therefore, we were motivated to analyze whether Auto-Sklearn 1's insight holds true for systems with higher-quality validation data. Consequently, we compared the performance of covariance matrix adaptation evolution strategy (CMA-ES), state-of-the-art gradient-free numerical optimization, to GES on the 71 classification datasets from the AutoML benchmark for AutoGluon. We found that Auto-Sklearn's insight depends on the chosen metric. For the metric ROC AUC, CMA-ES overfits drastically and is outperformed by GES -- statistically significantly for multi-class classification. For the metric balanced accuracy, CMA-ES does not overfit and outperforms GES significantly. Motivated by the successful application of CMA-ES for balanced accuracy, we explored methods to stop CMA-ES from overfitting for ROC AUC. We propose a method to normalize the weights produced by CMA-ES, inspired by GES, that avoids overfitting for CMA-ES and makes CMA-ES perform better than or similar to GES for ROC AUC.
- Biedrzycki, R. (2020). Handling bound constraints in CMA-ES: an experimental study. Swarm Evol. Comput., 52.
- Getting the most out of ensemble selection. In Proceedings of the 6th IEEE International Conference on Data Mining (ICDM 2006), 18-22 December 2006, Hong Kong, China, pages 828–833. IEEE Computer Society.
- Ensemble selection from libraries of models. In Brodley, C. E., editor, Machine Learning, Proceedings of the Twenty-first International Conference (ICML 2004), Banff, Alberta, Canada, July 4-8, 2004, volume 69 of ACM International Conference Proceeding Series. ACM.
- Autogluon-tabular: Robust and accurate automl for structured data. CoRR, abs/2003.06505.
- Efficient and robust automated machine learning. In Cortes, C., Lawrence, N. D., Lee, D. D., Sugiyama, M., and Garnett, R., editors, Advances in Neural Information Processing Systems 28: Annual Conference on Neural Information Processing Systems 2015, December 7-12, 2015, Montreal, Quebec, Canada, pages 2962–2970.
- AMLB: an automl benchmark. CoRR, abs/2207.12560.
- Hansen, B. E. (2007). Least squares model averaging. Econometrica, 75(4):1175–1189.
- Hansen, B. E. (2008). Least-squares forecast averaging. Journal of Econometrics, 146(2):342–350.
- Hansen, N. (2016). The CMA evolution strategy: A tutorial. CoRR, abs/1604.00772.
- CMA-ES/pycma on Github. Zenodo, DOI:10.5281/zenodo.2559634.
- Principled design of continuous stochastic search: From theory to practice. In Borenstein, Y. and Moraglio, A., editors, Theory and Principled Methods for the Design of Metaheuristics, Natural Computing Series, pages 145–180. Springer.
- Comparing results of 31 algorithms from the black-box optimization benchmarking BBOB-2009. In Pelikan, M. and Branke, J., editors, Genetic and Evolutionary Computation Conference, GECCO 2010, Proceedings, Portland, Oregon, USA, July 7-11, 2010, Companion Material, pages 1689–1696. ACM.
- Herbold, S. (2020). Autorank: A python package for automated ranking of classifiers. Journal of Open Source Software, 5(48):2173.
- Time efficiency in optimization with a bayesian-evolutionary algorithm. Swarm Evol. Comput., 69:100970.
- Model averaging is asymptotically better than model selection for prediction. J. Mach. Learn. Res., 23:33:1–33:53.
- H2O AutoML: Scalable automatic machine learning. 7th ICML Workshop on Automated Machine Learning (AutoML).
- Evolution strategies for continuous optimization: A survey of the state-of-the-art. Swarm and Evolutionary Computation, 56:100694.
- The algorithm selection competitions 2015 and 2017. Artif. Intell., 272:86–100.
- CMA-ES for hyperparameter optimization of deep neural networks. CoRR, abs/1604.07269.
- Towards automatically-tuned deep neural networks. In Hutter, F., Kotthoff, L., and Vanschoren, J., editors, AutoML: Methods, Sytems, Challenges, chapter 7, pages 141–156. Springer.
- Warm starting CMA-ES for hyperparameter optimization. In Thirty-Fifth AAAI Conference on Artificial Intelligence, AAAI 2021, Thirty-Third Conference on Innovative Applications of Artificial Intelligence, IAAI 2021, The Eleventh Symposium on Educational Advances in Artificial Intelligence, EAAI 2021, Virtual Event, February 2-9, 2021, pages 9188–9196. AAAI Press.
- Scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12:2825–2830.
- Mljar: State-of-the-art automated machine learning framework for tabular data. version 0.10.3.
- Assembled-openML: Creating efficient benchmarks for ensembles in autoML with openML. In First Conference on Automated Machine Learning (Late-Breaking Workshop).
- Szynkiewicz, P. (2018). Comparative study of pso and cma-es algorithms on black-box optimization benchmarks. Journal of Telecommunications and Information Technology.
- Development of a novel feature weighting method using CMA-ES optimization. In 26th Signal Processing and Communications Applications Conference, SIU 2018, Izmir, Turkey, May 2-5, 2018, pages 1–4. IEEE.
- An ensemble pruning primer. In Okun, O. and Valentini, G., editors, Applications of Supervised and Unsupervised Ensemble Methods, volume 245 of Studies in Computational Intelligence, pages 1–13. Springer.
- Super learner. Statistical Applications in Genetics and Molecular Biology, 6(1).
- Wolpert, D. H. (1992). Stacked generalization. Neural Networks, 5(2):241–259.
- Auto-pytorch tabular: Multi-fidelity metalearning for efficient and robust autodl. IEEE Transactions on Pattern Analysis and Machine Intelligence, pages 3079 – 3090. also available under https://arxiv.org/abs/2006.13799.