Towards Green Automated Machine Learning: Status Quo and Future Directions (2111.05850v4)
Abstract: Automated machine learning (AutoML) strives for the automatic configuration of machine learning algorithms and their composition into an overall (software) solution - a machine learning pipeline - tailored to the learning task (dataset) at hand. Over the last decade, AutoML has developed into an independent research field with hundreds of contributions. At the same time, AutoML is being criticised for its high resource consumption as many approaches rely on the (costly) evaluation of many machine learning pipelines, as well as the expensive large scale experiments across many datasets and approaches. In the spirit of recent work on Green AI, this paper proposes Green AutoML, a paradigm to make the whole AutoML process more environmentally friendly. Therefore, we first elaborate on how to quantify the environmental footprint of an AutoML tool. Afterward, different strategies on how to design and benchmark an AutoML tool wrt. their "greenness", i.e. sustainability, are summarized. Finally, we elaborate on how to be transparent about the environmental footprint and what kind of research incentives could direct the community into a more sustainable AutoML research direction. Additionally, we propose a sustainability checklist to be attached to every AutoML paper featuring all core aspects of Green AutoML.
- C. Ansótegui, M. Sellmann and K. Tierney “A Gender-Based Genetic Algorithm for the Automatic Configuration of Algorithms” In Proceedings of 15th International Conference on Principles and Practice of Constraint Programming (CP’09), 2009
- L. Wolff Anthony, B. Kanding and R. Selvan “Carbontracker: Tracking and Predicting the Carbon Footprint of Training Deep Learning Models” In arXiv/2007.03051, 2020
- N. Awad, N. Mallik and F. Hutter “DEHB: Evolutionary Hyberband for Scalable, Robust and Efficient Hyperparameter Optimization” In Proceedings of the 13th International Joint Conference on Artificial Intelligence (IJCAI’21), 2021
- “Benchmarking Automatic Machine Learning Frameworks” In arXiv/1808.06492, 2018
- “Collaborative Hyperparameter Tuning” In Proceedings of the 30th International Conference on Machine Learning (ICML’13), 2013
- “On the Dangers of Stochastic Parrots: Can Language Models be too big?” In Proceedings of the 4th ACM Conference on Fairness, Accountability, and Transparency (ACM FAccT’21), 2021
- “Algorithms for Hyper-Parameter Optimization” In Proceedings of the 25th Conference on Neural Information Processing Systems (NeurIPS’11), 2011
- L. Bliek “A Survey on Sustainable Surrogate-Based Optimisation” In Sustainability, 2022
- C. Calero, M. Bertoa and M. Á. Moraga “A Systematic Literature Review for Software Sustainability Measures” In GREENS 2013: 2nd International Workshop on Green and Sustainable Software, 2013
- “5Ws of Green and Sustainable Software” In Tsinghua Science and Technology, 2019
- A. Candelieri, R. Perego and F. Archetti “Green Machine Learning via Augmented Gaussian Processes and Multi-Information Source Optimization” In Soft Computing, 2021
- A. Candelieri, A. Ponti and F. Archetti “Fair and Green Hyperparameter Optimization via Multi-objective and Multiple Information Source Bayesian Optimization” In arXiv/2205.08835, 2022
- “IrEne: Interpretable Energy Prediction for Transformers” In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, (ACL/IJCNLP’21), 2021
- L. Cheng, K. Varshney and H. Liu “Socially Responsible AI Algorithms: Issues, Purposes, and Challenges” In Journal of Artificial Intelligence Research (JAIR’21), 2021
- T. Domhan, J. Springenberg and F. Hutter “Speeding up Automatic Hyperparameter Optimization of Deep Neural Networks by Extrapolation of Learning Curves” In Proceedings of the 24th International Joint Conference on Artificial Intelligence (IJCAI’15), 2015
- “NAS-Bench-201: Extending the Scope of Reproducible Neural Architecture Search” In Proceedings of the 8th International Conference on Learning Representations, (ICLR’20), 2020
- “AutoML using Metadata Language Embeddings” In arXiv/1910.03698, 2019
- “Neural Model-Based Optimization with Right-Censored Observations” In arXiv/2009.13828, 2020
- “MetaDL Challenge Design and Baseline Results” In AAAI: Workshop on Meta-Learning and MetaDL Challenge, 2021
- T. Elsken, J. Metzen and F. Hutter “Efficient Multi-Objective Neural Architecture Search via Lamarckian Evolution” In Proceedings of the 7th International Conference on Learning Representations (ICLR’19), 2019
- T. Elsken, J. Metzen and F. Hutter “Neural Architecture Search: A Survey” In Journal of Machine Learning Research (JMLR’19), 2019
- S. Falkner, A. Klein and F. Hutter “BOHB: Robust and Efficient Hyperparameter Optimization at Scale” In Proceedings of the 35th International Conference on Machine Learning (ICML’18), 2018
- “Practical Automated Machine Learning for the AutoML Challenge 2018” In ICML: International Workshop on Automatic Machine Learning, 2018
- “Auto-sklearn 2.0: Hands-Free AutoML via Meta-Learning” In Journal of Machine Learning Research (JMLR’22), 2022
- “Efficient and Robust Automated Machine Learning” In Proceedings of the 28th Conference on Neural Information Processing Systems (NeurIPS’15), 2015
- N. Fusi, R. Sheth and M. Elibol “Probabilistic Matrix Factorization for Automated Machine Learning” In Proceedings of the 31th Conference on Neural Information Processing Systems (NeurIPS’18), 2018
- “Estimation of Energy Consumption in Machine Learning” In Journal of Parallel and Distributed Computing, 2019
- M. Gendreau “An Introduction to Tabu Search” In Handbook of Metaheuristics, International Series in Operations Research & Management Science Springer, 2003
- “An Open Source AutoML Benchmark” In arXiv/1907.00909, 2019
- “Analysis of the AutoML Challenge Series 2015-2018” In Automated Machine Learning - Methods, Systems, Challenges, Springer Series on Challenges in Machine Learning Springer, 2019, pp. 177–219
- “AMC: AutoML for Model Compression and Acceleration on Mobile Devices” In Proceedings of the 15th European Conference on Computer Vision (ECCV’18), 2018
- “Predicting Execution Time of Computer Programs Using Sparse Polynomial Regression” In Proceedings of the 24th Conference on Neural Information Processing Systems (NeurIPS’10), 2010
- F. Hutter, H. Hoos and K. Leyton-Brown “Sequential Model-Based Optimization for General Algorithm Configuration” In Proceedings of the 5th International Conference on Learning and Intelligent Optimization (LION’11), 2011
- “ParamILS: An Automatic Algorithm Configuration Framework” In Journal of Artificial Intelligence Research (JAIR’09), 2009
- “Automated Machine Learning: Methods, Systems, Challenges” Springer, 2019
- “Algorithm Runtime Prediction: Methods & Evaluation” In Artificial Intelligence (AIJ’14), 2014
- “Non-Stochastic Best Arm Identification and Hyperparameter Optimization” In Proceedings of the 19th International Conference on Artificial Intelligence and Statistics (AISTATS’16), 2016
- “Multi-Fidelity Bayesian Optimisation with Continuous Approximations” In Proceedings of the 34th International Conference on Machine Learning (ICML’17), 2017
- “Fast Bayesian Optimization of Machine Learning Hyperparameters on Large Datasets” In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics (AISTATS’17), 2017
- “NAS-Bench-NLP: Neural Architecture Search Benchmark for Natural Language Processing” In IEEE Access, 2022
- B. Komer, J. Bergstra and C. Eliasmith “Hyperopt-Sklearn: Automatic Hyperparameter Configuration for scikit-learn” In ICML: Workshop on AutoML, 2014
- “Quantifying the Carbon Emissions of Machine Learning” In arXiv/1910.09700, 2019
- “Framing Sustainability as a Property of Software Quality” In Communications of the ACM, 2015
- “One Shot Learning of Simple Visual Concepts” In Proceedings of the 33th Annual Meeting of the Cognitive Science Society (CogSci’11), 2011
- L. Lannelongue, J. Grealey and M. Inouye “Green Algorithms: Quantifying the Carbon Footprint of Computation” In Advanced Science Wiley Online Library, 2021
- “HW-NAS-Bench: Hardware-Aware Neural Architecture Search Benchmark” In Proceedings of the 9th International Conference on Learning Representations (ICLR’21), 2021
- “Hyperband: A Novel Bandit-Based Approach to Hyperparameter Optimization” In Journal of Machine Learning Research (JMLR’17), 2017
- “Zen-NAS: A Zero-Shot NAS for High-Performance Deep Image Recognition” In arXiv/2102.01063, 2021
- “Warmstarting of Model-Based Algorithm Configuration” In Proceedings of the 32th Conference on Artificial Intelligence, (AAAI’18), 2018
- “Progressive Neural Architecture Search” In Proceedings of the 15th European Conference on Computer Vision (ECCV’18), 2018
- H. Liu, K. Simonyan and Y. Yang “DARTS: Differentiable Architecture Search” In Proceedings of the 7th International Conference on Learning Representations (ICLR’19), 2019
- “Automatic Termination for Hyperparameter Optimization” In Proceedings of the 1st Conference on Automated Machine Learning (AutoML’22), 2022
- “Engineering Design Optimization” Cambridge University Press, 2021
- “Neural Architecture Search without Training” In Proceedings of the 38th International Conference on Machine Learning (ICML’21), 2021
- “Towards Model Selection using Learning Curve Cross-Validation” In ICML: Workshop on Automated Machine Learning, 2021
- F. Mohr, M. Wever and E. Hüllermeier “ML-Plan: Automated Machine Learning via Hierarchical Planning” In Machine Learning (MLJ’18), 2018
- “Predicting Machine Learning Pipeline Runtimes in the Context of Automated Machine Learning” In IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI’21), 2021
- P. Nguyen, A. Kalousis and M. Hilario “A Meta-Mining Infrastructure to Support KD Workflow Optimization” In Proceedings of the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML/PKDD’11), 2011
- “The Energy and Carbon Footprint of Training End-to-End Speech Recognizers” In Proceedings of the 22nd Annual Conference of the International Speech Communication Association (INTERSPEECH’21), 2021
- “The Carbon Footprint of Machine Learning Training will Plateau, then Shrink” In Computer, 2022
- “Carbon Emissions and Large Neural Network Training” In arXiv/2104.10350, 2021
- “Multiple Adaptive Bayesian Linear Regression for Scalable Bayesian Optimization with Warm Start” In arXiv/1712.02902, 2017
- J. Petrak “Fast Subsampling Performance Estimates for Classification Algorithm Selection” In ECML: Workshop on Meta-Learning: Building Automatic Advice Strategies for Model Selection and Method Combination, 2000
- “Multi-Objective Automatic Machine Learning with AutoxgboostMC” In arXiv/1908.10796, 2019
- L. Prechelt “Early Stopping - But When?” In Neural Networks: Tricks of the Trade - Second Edition Springer, 2012
- “Regularized Evolution for Image Classifier Architecture Search” In Proceedings of the 33rd Conference on Artificial Intelligence (AAAI’19), 2019
- J. Reiner, R. Vaziri and N. Zobeiry “Machine Learning Assisted Characterisation and Simulation of Compressive Damage in Composite Laminates” In Composite Structures, 2021
- “Characterizing Classification Datasets: a Study of Meta-Features for Meta-Learning” In arXiv/1808.10406, 2018
- Y. Samo “LeanML: A Design Pattern To Slash Avoidable Wastes in Machine Learning Projects” In arXiv/2107.08066, 2021
- “CodeCarbon: Estimate and Track Carbon Emissions from Machine Learning Computing” In arXiv/2007.04074, 2021
- “Multi-Objective Asynchronous Successive Halving” In arXiv/2106.12639, 2021
- “Green AI” In arXiv/1907.10597, 2019
- “EnergyVis: Interactively Tracking and Exploring Energy Consumption for ML Models” In Proceedings of the Conference on Human Factors in Computing Systems (CHI’21), 2021
- “NAS-Bench-301 and the Case for Surrogate Benchmarks for Neural Architecture Search” In arXiv/2008.09777, 2020
- “Privileged Zero-Shot AutoML” In arXiv/2106.13743, 2021
- “Discovering the Suitability of Optimisation Algorithms by Learning from Evolved Instances” In Annals of Mathematics and Artificial Intelligence, 2011
- J. Snoek, H. Larochelle and R. Adams “Practical Bayesian Optimization of Machine Learning Algorithms” In Proceedings of the 26th Conference on Neural Information Processing Systems (NeurIPS’12), 2012
- “HyperPower: Power and Memory Constrained Hyperparameter Optimization for Neural Networks” In Proceedings of the Conference on Design, Automation and Test in Europe (DATE’18), 2018
- “Designing Adaptive Neural Networks for Energy-Constrained Image Classification” In Proceedings of the International Conference on Computer-Aided Design (ICCAD’18), 2018
- K. Swersky, J. Snoek and R. Adams “Multi-Task Bayesian Optimization” In Proceedings of the 27th Conference on Neural Information Processing Systems (NeurIPS’13), 2013
- K. Swersky, J. Snoek and R. Adams “Freeze-Thaw Bayesian Optimization” In arXiv/1406.3896, 2014
- “Auto-WEKA: Combined Selection and Hyperparameter Optimization of Classification Algorithms” In Proceedings of the 19th ACM International Conference on Knowledge Discovery and Data Mining, (KDD’13), 2013
- “Run2Survive: A Decision-Theoretic Approach to Algorithm Selection based on Survival Analysis” In Proceedings of the 12th Asian Conference on Machine Learning, (ACML’20), 2020
- “Towards Green Automated Machine Learning: Status Quo and Future Directions” In Journal of Artificial Intelligence Research 77, 2023, pp. 427–457
- “Coevolution of Remaining Useful Lifetime Estimation Pipelines for Automated Predictive Maintenance” In Proceedings of the Genetic and Evolutionary Computation Conference (GECCO’21), 2021
- “AutoML for Predictive Maintenance: One Tool to RUL Them All” In ECML/PKDD: Workshop on IoT Streams for Data-Driven Predictive Maintenance and IoT, Edge, and Mobile for Embedded Machine Learning, 2020
- “Transfer Learning” In Handbook of Research on Machine Learning Applications and Trends: Algorithms, Methods, and Techniques IGI global, 2010
- “AutoML for Climate Change: A Call to Action” In NeurIPS: Workshop on Tackling Climate Change with Machine Learning, 2022
- J. Vanschoren “Meta-Learning: A Survey” In arXiv/1810.03548, 2018
- “OpenML: Networked Science in Machine Learning” In SIGKDD Explorations, 2013
- “Automated Machine Learning for Short-term Electric Load Forecasting” In Proceedings of the IEEE Symposium Series on Computational Intelligence (SSCI’19), 2019
- “Energy-Aware Neural Architecture Optimization with Fast Splitting Steepest Descent” In arXiv/1910.03103, 2019
- “Generalizing from a Few Examples: A Survey on Few-Shot Learning” In ACM Computing Surveys, 2020
- “AutoML for Multi-Label Classification: Overview and Empirical Evaluation” In IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI’21), 2021
- “The FAIR Guiding Principles for Scientific Data Management and Stewardship” In Scientific Data, 2016
- “Practical Multi-Fidelity Bayesian Optimization for Hyperparameter Tuning” In Proceedings of the 35th Conference on Uncertainty in Artificial Intelligence (UAI’19), 2019
- A. Wynsberghe “Sustainable AI: AI for Sustainability and the Sustainability of AI” In AI Ethics, 2021
- Y. Xian, B. Schiele and Z. Akata “Zero-Shot Learning - The Good, the Bad and the Ugly” In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR’17), 2017
- “OBOE: Collaborative Filtering for AutoML Model Selection” In Proceedings of the 25th ACM International Conference on Knowledge Discovery and Data Mining, (KDD’19), 2019
- “NAS-Bench-101: Towards Reproducible Neural Architecture Search” In Proceedings of the 36th International Conference on Machine Learning (ICML’19), 2019
- “Efficient Transfer Learning Method for Automatic Hyperparameter Tuning” In Proceedings of the 17th International Conference on Artificial Intelligence and Statistics (AISTATS’14), 2014
- A. Zela, J. Siems and F. Hutter “NAS-Bench-1Shot1: Benchmarking and Dissecting One-Shot Neural Architecture Search” In Proceedings of the 8th International Conference on Learning Representations (ICLR’20), 2020
- L. Zimmer, M. Lindauer and F. Hutter “Auto-Pytorch: Multi-Fidelity Meta Learning for Efficient and Robust AutoDL” In IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI’21), 2021
- “Benchmark and Survey of Automated Machine Learning Frameworks” In Journal of Artificial Intelligence Research (JAIR’21), 2021
- “Neural Architecture Search with Reinforcement Learning” In Proceedings of the 5th International Conference on Learning Representations (ICLR’17), 2017
- Tanja Tornede (6 papers)
- Alexander Tornede (13 papers)
- Jonas Hanselle (4 papers)
- Marcel Wever (23 papers)
- Felix Mohr (18 papers)
- Eyke Hüllermeier (129 papers)