AIs rely on data centres that use vast amounts of energy Jason Alden/Bloomberg/Getty
Being more judicious in which AI models we use for tasks could potentially save 31.9 terawatt-hours of energy this year alone 鈥 equivalent to the output of five nuclear reactors.
at the University of Cote d鈥橝zur in France and his colleagues looked at 14 different tasks that people use generative AI tools for, ranging from text generation to speech recognition and image classification.
They then examined public leaderboards, including those hosted by the machine learning hub Hugging Face, for how different models perform. The energy efficiency of the models during inference 鈥 when an AI model produces an answer 鈥 was measured by a tool called CarbonTracker, and the total energy use of that model was calculated by tracking user downloads.
鈥淏ased on the size of the model, we estimated the energy consumption, and based on this, we can try to do our estimations,鈥 says da Silva Barros.
The researchers found that, across all 14 tasks, switching from the best-performing to the most energy-efficient models for each task reduced energy use by 65.8 per cent, while only making the output 3.9 per cent less useful 鈥 a trade-off they suggest could be acceptable to the public.
Free newsletter
Sign up to The Daily
The latest on what鈥檚 new in science and why it matters each day.

Because some people already use the most economical models, if people in the real world swapped from high-performance models to the most energy-efficient model they could bring about a 27.8 per cent reduction in energy consumption overall. 鈥淲e were surprised by how much can be saved,鈥 says team member at the French National Centre for Scientific Research.
However, that would require change from both users and AI companies, says da Silva Barros. 鈥淲e have to think in the direction of running small models, even if we lose some of the performance,鈥 he says. 鈥淎nd companies, when they develop models, it鈥檚 important they share some information on the model which allows the users to understand and evaluate if the model is very energy consuming or not.鈥
Some AI companies are reducing the energy consumption of their products through a process called model distillation, where large models are used to train smaller models. This is already having a significant impact, says at the University of Bristol in the UK. For example, Google recently claimed in Gemini over the past year.
However, getting users to pick the most efficient models 鈥渋s unlikely to result in limiting the energy increase from聽data centres as the authors suggest, at least in the current AI bubble,鈥 says Preist. 鈥淩educing energy per prompt will simply allow more customers to be served more rapidly with more sophisticated reasoning options,鈥 he says.
鈥淯sing smaller models can definitely result in less energy usage in the short term, but there are so many other factors that need to be considered when making any kind of meaningful projections into the future,鈥 says at Hugging Face. She cautions that rebound effects like increased use 鈥渉ave to be taken into account, as well as the broader impacts on society and the economy鈥.
Luccioni points out that any research in this space relies on external estimates and analysis because of a lack of transparency from individual companies. 鈥淲hat we need, to do these kinds of more complex analyses, is more transparency from AI companies, data centre operators and even governments,鈥 she says. 鈥淭his will allow researchers and policy-makers to make informed projections and decisions.鈥
Reference:
Topics:



