Concepts de base
The author investigates the energy consumption of deep neural networks, revealing non-linear relationships between network parameters, FLOPs, and energy use. The study emphasizes the impact of cache effects on energy efficiency.
Résumé
The study explores the complex relationship between dataset size, network structure, and energy consumption in deep neural networks. It introduces the BUTTER-E dataset, highlighting the surprising non-linear relationship between energy efficiency and network design. The analysis uncovers the critical role of cache-considerate algorithm development in achieving greater energy efficiency.
Large-scale neural networks contribute significantly to rising energy consumption in computing systems. Despite advancements in chip technology, AI's exponential growth leads to increased energy usage. The study proposes practical guidance for creating more energy-efficient neural networks and promoting sustainable AI.
Key findings include the impact of hyperparameter choice on energy efficiency across different architectures and counter-intuitive results related to hardware-mediated interactions. The research suggests a combined approach to developing energy-efficient architectures, algorithms, and hardware design.
Stats
One projection forecasts a substantial 20.9% proportion of the world’s total electricity demand will go to computing by 2030.
Training state-of-the-art AI models continue to double in both CO2 emissions and energy consumption every four to six months.
The BUTTER-E dataset contains data from 63,527 individual experimental runs spanning various configurations.
CPU-based training consumed a median marginal cost of 6.16mJ/datum with an upper quartile (UQ) of 16.32mJ/datum.
GPU-based training consumed a median marginal cost of 9.47mJ/datum with an upper quartile (UQ) of 13.75mJ/datum.
Citations
"Without substantial advances in “GreenAI” technologies to counter this “RedAI” trend, we are on course to dive head-first into troubling waters."
"Some may challenge this statement on the basis that computing hardware will become increasingly efficient."
"The urgent need to address AI’s energy efficiency has also been raised by the wider computer science community."