Assembled-OpenML: Creating Efficient Benchmarks for Ensembles in AutoML with OpenML (2307.00285v1)
Abstract: Automated Machine Learning (AutoML) frameworks regularly use ensembles. Developers need to compare different ensemble techniques to select appropriate techniques for an AutoML framework from the many potential techniques. So far, the comparison of ensemble techniques is often computationally expensive, because many base models must be trained and evaluated one or multiple times. Therefore, we present Assembled-OpenML. Assembled-OpenML is a Python tool, which builds meta-datasets for ensembles using OpenML. A meta-dataset, called Metatask, consists of the data of an OpenML task, the task's dataset, and prediction data from model evaluations for the task. We can make the comparison of ensemble techniques computationally cheaper by using the predictions stored in a metatask instead of training and evaluating base models. To introduce Assembled-OpenML, we describe the first version of our tool. Moreover, we present an example of using Assembled-OpenML to compare a set of ensemble techniques. For this example comparison, we built a benchmark using Assembled-OpenML and implemented ensemble techniques expecting predictions instead of base models as input. In our example comparison, we gathered the prediction data of $1523$ base models for $31$ datasets. Obtaining the prediction data for all base models using Assembled-OpenML took ${\sim} 1$ hour in total. In comparison, obtaining the prediction data by training and evaluating just one base model on the most computationally expensive dataset took ${\sim} 37$ minutes.
- Discrete neighborhood representations and modified stacked generalization methods for distributed regression. J. Univers. Comput. Sci., 21(6):842–855.
- Classifier subset selection for the stacked generalization method applied to emotion recognition in speech. Sensors, 16(1):21.
- Ensemble diversity measures and their application to thinning. Information Fusion, 6(1):49–62.
- On the predictive power of meta-features in openml. International Journal of Applied Mathematics and Computer Science, 27(4):697–712.
- Openml benchmarking suites. In Vanschoren, J. and Yeung, S., editors, Proceedings of the Neural Information Processing Systems Track on Datasets and Benchmarks 1, NeurIPS Datasets and Benchmarks 2021, December 2021, virtual.
- Machine learning for first-order theorem proving. Journal of automated reasoning, 53(2):141–172.
- Bulloch, B. et al. (1991). Eucalyptus species selection for soil conservation in seasonally dry hill country-twelfth year assessment. New Zealand journal of forestry science, 21(1):10–31.
- Ensemble selection from libraries of models. In Proceedings of the twenty-first international conference on Machine learning, page 18.
- Autostacker: a compositional evolutionary learning system. In Aguirre, H. E. and Takadama, K., editors, Proceedings of the Genetic and Evolutionary Computation Conference, GECCO 2018, Kyoto, Japan, July 15-19, 2018, pages 402–409. ACM.
- Ensemble machine learning-based algorithm for electric vehicle user behavior prediction. Applied Energy, 254:113732.
- Dynamic classifier selection: Recent advances and perspectives. Information Fusion, 41:195–216.
- Dietterich, T. G. (2000). Ensemble methods in machine learning. In International workshop on multiple classifier systems, pages 1–15. Springer.
- UCI machine learning repository.
- Efficient benchmarking of hyperparameter optimizers via surrogates. In Bonet, B. and Koenig, S., editors, Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence, January 25-30, 2015, Austin, Texas, USA, pages 1114–1120. AAAI Press.
- Autogluon-tabular: Robust and accurate automl for structured data. arXiv preprint arXiv:2003.06505.
- Overcoming process delays with decision tree induction. IEEE expert, 9(1):60–66.
- Efficient and robust automated machine learning. In Advances in Neural Information Processing Systems 28 (2015), pages 2962–2970.
- Openml-python: an extensible python api for openml. arXiv:1911.02490.
- Dynamic classifier selection based on multiple classifier behaviour. Pattern Recognition, 34(9):1879–1882.
- Ensemble machine learning models for the detection of energy theft. Electric Power Systems Research, 192:106904.
- Result analysis of the nips 2003 feature selection challenge. Advances in neural information processing systems, 17.
- An ensemble machine learning approach through effective feature extraction to classify fake news. Future Generation Computer Systems, 117:47–58.
- Array programming with NumPy. Nature, 585(7825):357–362.
- Algorithm runtime prediction: Methods & evaluation. Artif. Intell., 206:79–111.
- Caffe: Convolutional architecture for fast feature embedding. arXiv preprint arXiv:1408.5093.
- Automated algorithm selection: Survey and perspectives. Evol. Comput., 27(1):3–45.
- On combining classifiers. IEEE transactions on pattern analysis and machine intelligence, 20(3):226–239.
- Tabular benchmarks for joint architecture and hyperparameter optimization. arXiv preprint arXiv:1905.04970.
- From dynamic classifier selection to dynamic ensemble selection. Pattern recognition, 41(5):1718–1731.
- Kohavi, R. et al. (1996). Scaling up the accuracy of naive-bayes classifiers: A decision-tree hybrid. In Kdd, volume 96, pages 202–207.
- Learning multiple layers of features from tiny images.
- Automatic exploration of machine learning experiments on openml. arXiv preprint arXiv:1806.10961.
- Open algorithm selection challenge 2017: Setup and scenarios. In Proceedings of the Open Algorithm Selection Challenge 2017, Brussels, Belgium, September 11-12, 2017, volume 79 of Proceedings of Machine Learning Research, pages 1–7. PMLR.
- Failure analysis of parameter-induced simulation crashes in climate models. Geoscientific Model Development, 6(4):1157–1171.
- Gesture unit segmentation using support vector machines: segmenting gestures from rest positions. In Proceedings of the 28th Annual ACM Symposium on Applied Computing, pages 46–52.
- Quantitative structure–activity relationship models for ready biodegradability of chemicals. Journal of chemical information and modeling, 53(4):867–878.
- A data-driven approach to predict the success of bank telemarketing. Decision Support Systems, 62:22–31.
- Transformative machine learning. arXiv preprint arXiv:1811.03392.
- pandas development team, T. (2020). pandas-dev/pandas: Pandas.
- Scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12:2825–2830.
- provided by Semeion, D. (2010). Research center of sciences of communication, via sersale 117, 00128, rome, italy.
- Rice, J. R. (1976). The algorithm selection problem. In Advances in computers, volume 15, pages 65–118. Elsevier.
- The PROMISE repository of software engineering databases. School of Information Technology and Engineering, University of Ottawa, Canada.
- Siebert, J. P. (1987). Vehicle recognition using rule based methods.
- Simonoff, J. S. (2003). Analyzing categorical data, volume 496. Springer.
- Extreme algorithm selection with dyadic feature representation. In International Conference on Discovery Science, pages 309–324. Springer.
- The online performance estimation framework: heterogeneous ensemble learning for data streams. Machine Learning, 107(1):149–176.
- Openml: networked science in machine learning. CoRR, abs/1407.7722.
- SciPy 1.0: Fundamental Algorithms for Scientific Computing in Python. Nature Methods, 17:261–272.
- Automatic frankensteining: Creating complex ensembles autonomously. In Chawla, N. V. and Wang, W., editors, Proceedings of the 2017 SIAM International Conference on Data Mining, Houston, Texas, USA, April 27-29, 2017, pages 741–749. SIAM.
- Wolpert, D. H. (1992). Stacked generalization. Neural Networks, 5(2):241–259.
- Knowledge discovery on rfm model using bernoulli sequence. Expert Systems with Applications, 36(3):5866–5871.
- Nas-bench-101: Towards reproducible neural architecture search. In Chaudhuri, K. and Salakhutdinov, R., editors, Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 9-15 June 2019, Long Beach, California, USA, volume 97 of Proceedings of Machine Learning Research, pages 7105–7114. PMLR.
- Tensorflow model garden. GitHub.
- Forecasting skewed biased stochastic ozone days: analyses, solutions and beyond. Knowledge and Information Systems, 14(3):299–326.
- Enhanced extreme learning machine with stacked generalization. In 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence), pages 1191–1198. IEEE.
- Zhao, Y. (2022). Autodes: Automl pipeline generation of classification with dynamic ensemble strategy selection. arXiv preprint arXiv:2201.00207.