Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Measuring the Energy Consumption and Efficiency of Deep Neural Networks: An Empirical Analysis and Design Recommendations (2403.08151v1)

Published 13 Mar 2024 in cs.LG, cs.AI, and cs.NE

Abstract: Addressing the so-called Red-AI'' trend of rising energy consumption by large-scale neural networks, this study investigates the actual energy consumption, as measured by node-level watt-meters, of training various fully connected neural network architectures. We introduce the BUTTER-E dataset, an augmentation to the BUTTER Empirical Deep Learning dataset, containing energy consumption and performance data from 63,527 individual experimental runs spanning 30,582 distinct configurations: 13 datasets, 20 sizes (number of trainable parameters), 8 networkshapes'', and 14 depths on both CPU and GPU hardware collected using node-level watt-meters. This dataset reveals the complex relationship between dataset size, network structure, and energy use, and highlights the impact of cache effects. We propose a straightforward and effective energy model that accounts for network size, computing, and memory hierarchy. Our analysis also uncovers a surprising, hardware-mediated non-linear relationship between energy efficiency and network design, challenging the assumption that reducing the number of parameters or FLOPs is the best way to achieve greater energy efficiency. Highlighting the need for cache-considerate algorithm development, we suggest a combined approach to energy efficient network, algorithm, and hardware design. This work contributes to the fields of sustainable computing and Green AI, offering practical guidance for creating more energy-efficient neural networks and promoting sustainable AI.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (35)
  1. U. S. Environmental Protection Agency. Emissions & Generation Resource Integrated Database (eGRID) Data Explorer. https://www.epa.gov/egrid/data-explorer, 2024. Updated: 2024-01-30.
  2. Anders S G Andrae. New perspectives on internet electricity use in 2030. 2020.
  3. EC-NAS: Energy Consumption Aware Tabular Benchmarks for Neural Architecture Search, May 2023. URL http://arxiv.org/abs/2210.06015. arXiv:2210.06015 [cs, stat].
  4. Exploring the Accuracy – Energy Trade-off in Machine Learning. In 2021 IEEE/ACM International Workshop on Genetic Improvement (GI), pages 11–18, Madrid, Spain, May 2021. IEEE. ISBN 978-1-66544-466-8. doi: 10.1109/GI52543.2021.00011. URL https://ieeexplore.ieee.org/document/9474356/.
  5. Neuralpower: Predict and deploy energy-efficient convolutional neural networks. In Asian Conference on Machine Learning, pages 622–637. PMLR, 2017.
  6. Cpu db: Recording microprocessor history: With this open database, you can mine microprocessor trends over the past 40 years. Queue, 10(4):10–27, 2012.
  7. Li Deng. The mnist database of handwritten digit images for machine learning research. IEEE Signal Processing Magazine, 29(6):141–142, 2012.
  8. Design of ion-implanted mosfet’s with very small physical dimensions. IEEE Journal of solid-state circuits, 9(5):256–268, 1974.
  9. Design of ion-implanted mosfet’s with very small physical dimensions. Proceedings of the IEEE, 87(4):668–678, 1999.
  10. Compute and energy consumption trends in deep learning inference. arXiv preprint arXiv:2109.05472, 2021.
  11. EpochAI. Parameter, compute and data trends in machine learning. https://epochai.org/data/pcd, 2023. Updated: 12/31/23.
  12. Estimation of energy consumption in machine learning. Journal of Parallel and Distributed Computing, 134:75–88, 2019.
  13. Nicola Jones et al. How to stop data centres from gobbling up the world’s electricity. Nature, 561(7722):163–166, 2018.
  14. Influence of random topology in artificial neural networks: A survey. ICT Express, 6(2):145–150, 2020.
  15. Adam: A method for stochastic optimization. In International Conference on Learning Representations (ICLR), San Diega, CA, USA, 2015.
  16. Implications of historical trends in the electrical efficiency of computing. IEEE Annals of the History of Computing, 33(3):46–54, 2011.
  17. A Transistor Operations Model for Deep Learning Energy Consumption Scaling, May 2022. URL http://arxiv.org/abs/2205.15062. Number: arXiv:2205.15062 arXiv:2205.15062 [cs].
  18. Shiwei Liu et al. Freetickets: Accurate, robust and efficient deep ensemble by training with dynamic sparsity. In Sparsity in Neural Networks: Advancing Understanding and Practice 2021, 2021.
  19. Igor L. Markov. Limits on fundamental limits to computation. Nature, 512(7513):147–154, 2014.
  20. A comparative analysis of artificial neural network architectures for building energy consumption forecasting. International Journal of Distributed Sensor Networks, 15(9):1550147719877616, 2019.
  21. Gordon E. Moore. Cramming more components onto integrated circuits. Electronics, 38(8):114–117, 1965.
  22. Pmlb: a large benchmark suite for machine learning evaluation and comparison. BioData Mining, 10(1):36, Dec 2017. ISSN 1756-0381. doi: 10.1186/s13040-017-0154-4. URL https://doi.org/10.1186/s13040-017-0154-4.
  23. The Energy and Carbon Footprint of Training End-to-End Speech Recognizers. In Interspeech 2021, pages 4583–4587. ISCA, August 2021. doi: 10.21437/Interspeech.2021-456. URL https://www.isca-speech.org/archive/interspeech_2021/parcollet21_interspeech.html.
  24. Synergy: An energy measurement and prediction framework for convolutional neural networks on jetson tx1. In Proceedings of the International Conference on Parallel and Distributed Processing Techniques and Applications (PDPTA), pages 375–382. The Steering Committee of The World Congress in Computer Science, Computer …, 2018.
  25. Tackling climate change with machine learning. arXiv, arXiv:1906.05433, 2019.
  26. Pmlb v1.0: an open source dataset collection for benchmarking machine learning methods. arXiv preprint arXiv:2012.00058v2, 2021.
  27. Green ai. Communications of the ACM, 63(12):54–63, 2020.
  28. Statista. What is the average annual power usage effectiveness (PUE) for your largest data center? https://www.statista.com/statistics/1229367/data-center-average-annual-pue-worldwide/, 2023.
  29. Energy and policy considerations for deep learning in nlp. arXiv, arXiv:1906.02243, 2019.
  30. TOP500 Project. TOP500 Supercomputer Sites. https://www.top500.org/, 2023. Accessed: 2024-02-01.
  31. Butter - empirical deep learning dataset. 5 2022a. doi: 10.25984/1872441.
  32. An empirical deep dive into deep learning’s driving dynamics. arXiv preprint arXiv:2207.12547, 2022b.
  33. Energy efficiency of training neural network architectures: An empirical study, 2023.
  34. A method to estimate the energy consumption of deep neural networks. In 2017 51st Asilomar Conference on Signals, Systems, and Computers, pages 1916–1920, October 2017. doi: 10.1109/ACSSC.2017.8335698. ISSN: 2576-2303.
  35. Nas-bench-101: Towards reproducible neural architecture search. In International conference on machine learning, pages 7105–7114. PMLR, 2019.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Charles Edison Tripp (2 papers)
  2. Jordan Perr-Sauer (2 papers)
  3. Jamil Gafur (3 papers)
  4. Amabarish Nag (1 paper)
  5. Avi Purkayastha (1 paper)
  6. Sagi Zisman (2 papers)
  7. Erik A. Bensen (3 papers)
Citations (1)
Youtube Logo Streamline Icon: https://streamlinehq.com