Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Towards Green Automated Machine Learning: Status Quo and Future Directions (2111.05850v4)

Published 10 Nov 2021 in cs.LG

Abstract: Automated machine learning (AutoML) strives for the automatic configuration of machine learning algorithms and their composition into an overall (software) solution - a machine learning pipeline - tailored to the learning task (dataset) at hand. Over the last decade, AutoML has developed into an independent research field with hundreds of contributions. At the same time, AutoML is being criticised for its high resource consumption as many approaches rely on the (costly) evaluation of many machine learning pipelines, as well as the expensive large scale experiments across many datasets and approaches. In the spirit of recent work on Green AI, this paper proposes Green AutoML, a paradigm to make the whole AutoML process more environmentally friendly. Therefore, we first elaborate on how to quantify the environmental footprint of an AutoML tool. Afterward, different strategies on how to design and benchmark an AutoML tool wrt. their "greenness", i.e. sustainability, are summarized. Finally, we elaborate on how to be transparent about the environmental footprint and what kind of research incentives could direct the community into a more sustainable AutoML research direction. Additionally, we propose a sustainability checklist to be attached to every AutoML paper featuring all core aspects of Green AutoML.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (105)
  1. C. Ansótegui, M. Sellmann and K. Tierney “A Gender-Based Genetic Algorithm for the Automatic Configuration of Algorithms” In Proceedings of 15th International Conference on Principles and Practice of Constraint Programming (CP’09), 2009
  2. L. Wolff Anthony, B. Kanding and R. Selvan “Carbontracker: Tracking and Predicting the Carbon Footprint of Training Deep Learning Models” In arXiv/2007.03051, 2020
  3. N. Awad, N. Mallik and F. Hutter “DEHB: Evolutionary Hyberband for Scalable, Robust and Efficient Hyperparameter Optimization” In Proceedings of the 13th International Joint Conference on Artificial Intelligence (IJCAI’21), 2021
  4. “Benchmarking Automatic Machine Learning Frameworks” In arXiv/1808.06492, 2018
  5. “Collaborative Hyperparameter Tuning” In Proceedings of the 30th International Conference on Machine Learning (ICML’13), 2013
  6. “On the Dangers of Stochastic Parrots: Can Language Models be too big?” In Proceedings of the 4th ACM Conference on Fairness, Accountability, and Transparency (ACM FAccT’21), 2021
  7. “Algorithms for Hyper-Parameter Optimization” In Proceedings of the 25th Conference on Neural Information Processing Systems (NeurIPS’11), 2011
  8. L. Bliek “A Survey on Sustainable Surrogate-Based Optimisation” In Sustainability, 2022
  9. C. Calero, M. Bertoa and M. Á. Moraga “A Systematic Literature Review for Software Sustainability Measures” In GREENS 2013: 2nd International Workshop on Green and Sustainable Software, 2013
  10. “5Ws of Green and Sustainable Software” In Tsinghua Science and Technology, 2019
  11. A. Candelieri, R. Perego and F. Archetti “Green Machine Learning via Augmented Gaussian Processes and Multi-Information Source Optimization” In Soft Computing, 2021
  12. A. Candelieri, A. Ponti and F. Archetti “Fair and Green Hyperparameter Optimization via Multi-objective and Multiple Information Source Bayesian Optimization” In arXiv/2205.08835, 2022
  13. “IrEne: Interpretable Energy Prediction for Transformers” In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, (ACL/IJCNLP’21), 2021
  14. L. Cheng, K. Varshney and H. Liu “Socially Responsible AI Algorithms: Issues, Purposes, and Challenges” In Journal of Artificial Intelligence Research (JAIR’21), 2021
  15. T. Domhan, J. Springenberg and F. Hutter “Speeding up Automatic Hyperparameter Optimization of Deep Neural Networks by Extrapolation of Learning Curves” In Proceedings of the 24th International Joint Conference on Artificial Intelligence (IJCAI’15), 2015
  16. “NAS-Bench-201: Extending the Scope of Reproducible Neural Architecture Search” In Proceedings of the 8th International Conference on Learning Representations, (ICLR’20), 2020
  17. “AutoML using Metadata Language Embeddings” In arXiv/1910.03698, 2019
  18. “Neural Model-Based Optimization with Right-Censored Observations” In arXiv/2009.13828, 2020
  19. “MetaDL Challenge Design and Baseline Results” In AAAI: Workshop on Meta-Learning and MetaDL Challenge, 2021
  20. T. Elsken, J. Metzen and F. Hutter “Efficient Multi-Objective Neural Architecture Search via Lamarckian Evolution” In Proceedings of the 7th International Conference on Learning Representations (ICLR’19), 2019
  21. T. Elsken, J. Metzen and F. Hutter “Neural Architecture Search: A Survey” In Journal of Machine Learning Research (JMLR’19), 2019
  22. S. Falkner, A. Klein and F. Hutter “BOHB: Robust and Efficient Hyperparameter Optimization at Scale” In Proceedings of the 35th International Conference on Machine Learning (ICML’18), 2018
  23. “Practical Automated Machine Learning for the AutoML Challenge 2018” In ICML: International Workshop on Automatic Machine Learning, 2018
  24. “Auto-sklearn 2.0: Hands-Free AutoML via Meta-Learning” In Journal of Machine Learning Research (JMLR’22), 2022
  25. “Efficient and Robust Automated Machine Learning” In Proceedings of the 28th Conference on Neural Information Processing Systems (NeurIPS’15), 2015
  26. N. Fusi, R. Sheth and M. Elibol “Probabilistic Matrix Factorization for Automated Machine Learning” In Proceedings of the 31th Conference on Neural Information Processing Systems (NeurIPS’18), 2018
  27. “Estimation of Energy Consumption in Machine Learning” In Journal of Parallel and Distributed Computing, 2019
  28. M. Gendreau “An Introduction to Tabu Search” In Handbook of Metaheuristics, International Series in Operations Research & Management Science Springer, 2003
  29. “An Open Source AutoML Benchmark” In arXiv/1907.00909, 2019
  30. “Analysis of the AutoML Challenge Series 2015-2018” In Automated Machine Learning - Methods, Systems, Challenges, Springer Series on Challenges in Machine Learning Springer, 2019, pp. 177–219
  31. “AMC: AutoML for Model Compression and Acceleration on Mobile Devices” In Proceedings of the 15th European Conference on Computer Vision (ECCV’18), 2018
  32. “Predicting Execution Time of Computer Programs Using Sparse Polynomial Regression” In Proceedings of the 24th Conference on Neural Information Processing Systems (NeurIPS’10), 2010
  33. F. Hutter, H. Hoos and K. Leyton-Brown “Sequential Model-Based Optimization for General Algorithm Configuration” In Proceedings of the 5th International Conference on Learning and Intelligent Optimization (LION’11), 2011
  34. “ParamILS: An Automatic Algorithm Configuration Framework” In Journal of Artificial Intelligence Research (JAIR’09), 2009
  35. “Automated Machine Learning: Methods, Systems, Challenges” Springer, 2019
  36. “Algorithm Runtime Prediction: Methods & Evaluation” In Artificial Intelligence (AIJ’14), 2014
  37. “Non-Stochastic Best Arm Identification and Hyperparameter Optimization” In Proceedings of the 19th International Conference on Artificial Intelligence and Statistics (AISTATS’16), 2016
  38. “Multi-Fidelity Bayesian Optimisation with Continuous Approximations” In Proceedings of the 34th International Conference on Machine Learning (ICML’17), 2017
  39. “Fast Bayesian Optimization of Machine Learning Hyperparameters on Large Datasets” In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics (AISTATS’17), 2017
  40. “NAS-Bench-NLP: Neural Architecture Search Benchmark for Natural Language Processing” In IEEE Access, 2022
  41. B. Komer, J. Bergstra and C. Eliasmith “Hyperopt-Sklearn: Automatic Hyperparameter Configuration for scikit-learn” In ICML: Workshop on AutoML, 2014
  42. “Quantifying the Carbon Emissions of Machine Learning” In arXiv/1910.09700, 2019
  43. “Framing Sustainability as a Property of Software Quality” In Communications of the ACM, 2015
  44. “One Shot Learning of Simple Visual Concepts” In Proceedings of the 33th Annual Meeting of the Cognitive Science Society (CogSci’11), 2011
  45. L. Lannelongue, J. Grealey and M. Inouye “Green Algorithms: Quantifying the Carbon Footprint of Computation” In Advanced Science Wiley Online Library, 2021
  46. “HW-NAS-Bench: Hardware-Aware Neural Architecture Search Benchmark” In Proceedings of the 9th International Conference on Learning Representations (ICLR’21), 2021
  47. “Hyperband: A Novel Bandit-Based Approach to Hyperparameter Optimization” In Journal of Machine Learning Research (JMLR’17), 2017
  48. “Zen-NAS: A Zero-Shot NAS for High-Performance Deep Image Recognition” In arXiv/2102.01063, 2021
  49. “Warmstarting of Model-Based Algorithm Configuration” In Proceedings of the 32th Conference on Artificial Intelligence, (AAAI’18), 2018
  50. “Progressive Neural Architecture Search” In Proceedings of the 15th European Conference on Computer Vision (ECCV’18), 2018
  51. H. Liu, K. Simonyan and Y. Yang “DARTS: Differentiable Architecture Search” In Proceedings of the 7th International Conference on Learning Representations (ICLR’19), 2019
  52. “Automatic Termination for Hyperparameter Optimization” In Proceedings of the 1st Conference on Automated Machine Learning (AutoML’22), 2022
  53. “Engineering Design Optimization” Cambridge University Press, 2021
  54. “Neural Architecture Search without Training” In Proceedings of the 38th International Conference on Machine Learning (ICML’21), 2021
  55. “Towards Model Selection using Learning Curve Cross-Validation” In ICML: Workshop on Automated Machine Learning, 2021
  56. F. Mohr, M. Wever and E. Hüllermeier “ML-Plan: Automated Machine Learning via Hierarchical Planning” In Machine Learning (MLJ’18), 2018
  57. “Predicting Machine Learning Pipeline Runtimes in the Context of Automated Machine Learning” In IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI’21), 2021
  58. P. Nguyen, A. Kalousis and M. Hilario “A Meta-Mining Infrastructure to Support KD Workflow Optimization” In Proceedings of the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML/PKDD’11), 2011
  59. “The Energy and Carbon Footprint of Training End-to-End Speech Recognizers” In Proceedings of the 22nd Annual Conference of the International Speech Communication Association (INTERSPEECH’21), 2021
  60. “The Carbon Footprint of Machine Learning Training will Plateau, then Shrink” In Computer, 2022
  61. “Carbon Emissions and Large Neural Network Training” In arXiv/2104.10350, 2021
  62. “Multiple Adaptive Bayesian Linear Regression for Scalable Bayesian Optimization with Warm Start” In arXiv/1712.02902, 2017
  63. J. Petrak “Fast Subsampling Performance Estimates for Classification Algorithm Selection” In ECML: Workshop on Meta-Learning: Building Automatic Advice Strategies for Model Selection and Method Combination, 2000
  64. “Multi-Objective Automatic Machine Learning with AutoxgboostMC” In arXiv/1908.10796, 2019
  65. L. Prechelt “Early Stopping - But When?” In Neural Networks: Tricks of the Trade - Second Edition Springer, 2012
  66. “Regularized Evolution for Image Classifier Architecture Search” In Proceedings of the 33rd Conference on Artificial Intelligence (AAAI’19), 2019
  67. J. Reiner, R. Vaziri and N. Zobeiry “Machine Learning Assisted Characterisation and Simulation of Compressive Damage in Composite Laminates” In Composite Structures, 2021
  68. “Characterizing Classification Datasets: a Study of Meta-Features for Meta-Learning” In arXiv/1808.10406, 2018
  69. Y. Samo “LeanML: A Design Pattern To Slash Avoidable Wastes in Machine Learning Projects” In arXiv/2107.08066, 2021
  70. “CodeCarbon: Estimate and Track Carbon Emissions from Machine Learning Computing” In arXiv/2007.04074, 2021
  71. “Multi-Objective Asynchronous Successive Halving” In arXiv/2106.12639, 2021
  72. “Green AI” In arXiv/1907.10597, 2019
  73. “EnergyVis: Interactively Tracking and Exploring Energy Consumption for ML Models” In Proceedings of the Conference on Human Factors in Computing Systems (CHI’21), 2021
  74. “NAS-Bench-301 and the Case for Surrogate Benchmarks for Neural Architecture Search” In arXiv/2008.09777, 2020
  75. “Privileged Zero-Shot AutoML” In arXiv/2106.13743, 2021
  76. “Discovering the Suitability of Optimisation Algorithms by Learning from Evolved Instances” In Annals of Mathematics and Artificial Intelligence, 2011
  77. J. Snoek, H. Larochelle and R. Adams “Practical Bayesian Optimization of Machine Learning Algorithms” In Proceedings of the 26th Conference on Neural Information Processing Systems (NeurIPS’12), 2012
  78. “HyperPower: Power and Memory Constrained Hyperparameter Optimization for Neural Networks” In Proceedings of the Conference on Design, Automation and Test in Europe (DATE’18), 2018
  79. “Designing Adaptive Neural Networks for Energy-Constrained Image Classification” In Proceedings of the International Conference on Computer-Aided Design (ICCAD’18), 2018
  80. K. Swersky, J. Snoek and R. Adams “Multi-Task Bayesian Optimization” In Proceedings of the 27th Conference on Neural Information Processing Systems (NeurIPS’13), 2013
  81. K. Swersky, J. Snoek and R. Adams “Freeze-Thaw Bayesian Optimization” In arXiv/1406.3896, 2014
  82. “Auto-WEKA: Combined Selection and Hyperparameter Optimization of Classification Algorithms” In Proceedings of the 19th ACM International Conference on Knowledge Discovery and Data Mining, (KDD’13), 2013
  83. “Run2Survive: A Decision-Theoretic Approach to Algorithm Selection based on Survival Analysis” In Proceedings of the 12th Asian Conference on Machine Learning, (ACML’20), 2020
  84. “Towards Green Automated Machine Learning: Status Quo and Future Directions” In Journal of Artificial Intelligence Research 77, 2023, pp. 427–457
  85. “Coevolution of Remaining Useful Lifetime Estimation Pipelines for Automated Predictive Maintenance” In Proceedings of the Genetic and Evolutionary Computation Conference (GECCO’21), 2021
  86. “AutoML for Predictive Maintenance: One Tool to RUL Them All” In ECML/PKDD: Workshop on IoT Streams for Data-Driven Predictive Maintenance and IoT, Edge, and Mobile for Embedded Machine Learning, 2020
  87. “Transfer Learning” In Handbook of Research on Machine Learning Applications and Trends: Algorithms, Methods, and Techniques IGI global, 2010
  88. “AutoML for Climate Change: A Call to Action” In NeurIPS: Workshop on Tackling Climate Change with Machine Learning, 2022
  89. J. Vanschoren “Meta-Learning: A Survey” In arXiv/1810.03548, 2018
  90. “OpenML: Networked Science in Machine Learning” In SIGKDD Explorations, 2013
  91. “Automated Machine Learning for Short-term Electric Load Forecasting” In Proceedings of the IEEE Symposium Series on Computational Intelligence (SSCI’19), 2019
  92. “Energy-Aware Neural Architecture Optimization with Fast Splitting Steepest Descent” In arXiv/1910.03103, 2019
  93. “Generalizing from a Few Examples: A Survey on Few-Shot Learning” In ACM Computing Surveys, 2020
  94. “AutoML for Multi-Label Classification: Overview and Empirical Evaluation” In IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI’21), 2021
  95. “The FAIR Guiding Principles for Scientific Data Management and Stewardship” In Scientific Data, 2016
  96. “Practical Multi-Fidelity Bayesian Optimization for Hyperparameter Tuning” In Proceedings of the 35th Conference on Uncertainty in Artificial Intelligence (UAI’19), 2019
  97. A. Wynsberghe “Sustainable AI: AI for Sustainability and the Sustainability of AI” In AI Ethics, 2021
  98. Y. Xian, B. Schiele and Z. Akata “Zero-Shot Learning - The Good, the Bad and the Ugly” In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR’17), 2017
  99. “OBOE: Collaborative Filtering for AutoML Model Selection” In Proceedings of the 25th ACM International Conference on Knowledge Discovery and Data Mining, (KDD’19), 2019
  100. “NAS-Bench-101: Towards Reproducible Neural Architecture Search” In Proceedings of the 36th International Conference on Machine Learning (ICML’19), 2019
  101. “Efficient Transfer Learning Method for Automatic Hyperparameter Tuning” In Proceedings of the 17th International Conference on Artificial Intelligence and Statistics (AISTATS’14), 2014
  102. A. Zela, J. Siems and F. Hutter “NAS-Bench-1Shot1: Benchmarking and Dissecting One-Shot Neural Architecture Search” In Proceedings of the 8th International Conference on Learning Representations (ICLR’20), 2020
  103. L. Zimmer, M. Lindauer and F. Hutter “Auto-Pytorch: Multi-Fidelity Meta Learning for Efficient and Robust AutoDL” In IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI’21), 2021
  104. “Benchmark and Survey of Automated Machine Learning Frameworks” In Journal of Artificial Intelligence Research (JAIR’21), 2021
  105. “Neural Architecture Search with Reinforcement Learning” In Proceedings of the 5th International Conference on Learning Representations (ICLR’17), 2017
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Tanja Tornede (6 papers)
  2. Alexander Tornede (13 papers)
  3. Jonas Hanselle (4 papers)
  4. Marcel Wever (23 papers)
  5. Felix Mohr (18 papers)
  6. Eyke Hüllermeier (129 papers)
Citations (32)