Can we make AI less power-hungry? These researchers are working on it.

Can we make AI less power-hungry? These researchers are working on it.

  • 24.03.2025 11:00
  • arstechnica.com
  • Keywords: AI, Energy Consumption, Data Centers, Nvidia, OpenAI, Google, Meta, AlexNet, ML Energy Initiative, GPU, Transformer Models, Quantization, Pruning, Perseus, ZeusMonitor, Moore’s Law, Photonic Chips, 2D Semiconductors

AI models' growing power consumption is driving researchers to optimize efficiency through techniques like pruning and quantization, while also exploring advanced hardware and data center management to reduce energy use. Despite progress, transparency from companies about AI energy demands remains a challenge.

Meta ReportsNvidia ServicesNvidia ReportsAlphabet ServicesMeta ServicesAMZNsentiment_dissatisfiedNVDAsentiment_satisfiedMETAsentiment_satisfied

Estimated market influence

Amazon

Amazon

Negativesentiment_dissatisfied
Analyst rating: Strong buy

FERC rejected Amazon's request to buy additional power directly from a nuclear plant for their data center, which could impact their operations and costs.

Nvidia

Nvidia

Positivesentiment_satisfied
Analyst rating: Strong buy

Nvidia's GPUs have become 15 times more efficient between 2010-2020, and 10-fold boost since 2020. Their chips are used in data centers for AI model training and inference.

OpenAI

Positivesentiment_satisfied
Analyst rating: N/A

Developed GPT models which have significantly increased AI's power consumption. OpenAI's lack of transparency on energy use hinders optimizations.

Google

Neutralsentiment_neutral
Analyst rating: N/A

Uses AI in their services, with potential future increases in power consumption if they integrate more LLMs. Their past statements on power usage are questioned for accuracy.

Meta

Meta

Positivesentiment_satisfied
Analyst rating: Strong buy

Developed the Llama model and contributed to AI efficiency through initiatives like ML Energy Initiative.

Electric Power Research Institute (EPRI)

Neutralsentiment_neutral
Analyst rating: N/A

Provided estimates on data center power consumption but no direct role in company operations or AI development.

US Federal Energy Regulatory Commission (FERC)

Negativesentiment_dissatisfied
Analyst rating: N/A

Rejected Amazon's power request, affecting the company's ability to manage energy costs for their data centers.

University of Toronto

Neutralsentiment_neutral
Analyst rating: N/A

Home to AI researchers who developed AlexNet, contributing to the rise in AI model complexity and power usage.

Lawrence Berkeley National Laboratory

Neutralsentiment_neutral
Analyst rating: N/A

Conducted studies on data center energy consumption but not directly involved in company operations or AI development beyond research.

University of Michigan

Neutralsentiment_neutral
Analyst rating: N/A

Hosted researchers contributing to AI efficiency, like Jae-Won Chung and Mosharaf Chowdhury, who are part of the ML Energy Initiative.

Context

Analysis of AI Power Consumption and Efficiency Innovations

Key Facts and Data Points

  • Amazon's Rejected Power Request:

    • Amazon sought an additional 180 megawatts (MW) of power directly from the Susquehanna nuclear plant for a nearby data center.
    • The request was rejected due to concerns it would disadvantage other grid users.
  • US Data Center Power Consumption:

    • Between 2018 and 2023, US data center power consumption rose from 76 TWh to 176 TWh.
    • Lawrence Berkeley National Laboratory estimates suggest annual energy draw in the US could reach 325-580 TWh by 2028, representing 6.7-12% of total US electricity consumption.
  • Historical Context:

    • The rise of AI power consumption began with AlexNet (2012), a CNN with over 60 million parameters and 650,000 neurons, marking the shift to GPU-based training.
    • Between 2010-2020, Nvidia's data center chips became 15 times more efficient.
  • AI Model Training:

    • Training GPT-4 reportedly used over 25,000 Nvidia Ampere 100 GPUs for 100 days, consuming 50 GW-hours of power (enough to power a medium-sized town for a year).
    • Google estimates that a single AI-powered search could require 400,000 new servers, consuming 22.8 TWh annually.
  • Energy Efficiency Innovations:

    • Techniques like pruning and quantization have reduced model size and power consumption.
    • Nvidia's quantization training cut memory requirements by 29-51%.
    • The Perseus tool, developed at the University of Michigan, reduced energy consumption in data centers by up to 30% during testing.
  • Proprietary Model Transparency:

    • Lack of transparency from companies like OpenAI and Google hinders benchmarking efforts.
    • Published estimates for AI power consumption are often based on outdated hardware or unverified metrics.
  • Future Technologies:

    • Advances in photonic chips (light-based computing) could reduce energy consumption by orders of magnitude.
    • 2D semiconductors and stacked transistors promise improved computation density and efficiency.

Market Trends and Business Impact

  • Shift to GPU-Based AI Training:

    • The transition from CPU to GPU, and now specialized hardware (e.g., TPUs), has driven significant increases in computational power but also energy consumption.
    • Nvidia's leadership in AI chips positions it as a key player in the race for efficiency.
  • Energy Efficiency as a Competitive Advantage:

    • Companies investing in energy-efficient AI technologies (e.g., pruning, quantization, and optimized data center operations) will gain a competitive edge in cost reduction and scalability.
    • OpenAI and Google's lack of transparency may slow collaborative innovation in this space.
  • Regulatory and Market Pressures:

    • Rising energy costs and environmental concerns are pushing governments and businesses to prioritize AI efficiency.
    • Potential regulatory scrutiny over data center power consumption could force companies to adopt greener practices.

Competitive Dynamics

  • Nvidia's Dominance:

    • Nvidia's leadership in GPU technology and its focus on improving energy efficiency have solidified its position in the AI hardware market.
    • Its advancements in chips like the H100 and software optimizations (e.g., quantization-aware training) are critical to maintaining competitiveness.
  • Emerging Technologies:

    • Startups and research labs focusing on alternative computing architectures (e.g., photonic chips) could disrupt the market if they achieve commercial viability.
    • Collaboration between academia and industry will be key to accelerating innovation.

Long-Term Effects and Regulatory Implications

  • Energy Paradox:

    • Improvements in AI efficiency may lead to increased adoption, potentially offsetting gains in energy savings.
    • This "rebound effect" could exacerbate the problem unless managed through policy or market mechanisms.
  • Potential for Regulatory Intervention:

    • Governments may impose stricter regulations on data center power consumption, forcing companies to adopt more sustainable practices.
    • Disclosure requirements for AI model power usage could become mandatory, fostering transparency and innovation.

Strategic Considerations

  • Investment in Efficiency Technologies:

    • Businesses should prioritize investments in energy-efficient AI technologies (e.g., optimized hardware, pruning tools) to reduce costs and improve sustainability.
    • Partnerships with research institutions like the ML Energy Initiative could yield significant benefits.
  • Transparency and Collaboration:

    • Open sharing of AI model power consumption data by companies like OpenAI and Google would accelerate innovation and foster trust among stakeholders.
    • Industry standards for benchmarking AI energy efficiency could drive collective progress.

Conclusion

The race to make AI less power-hungry is critical for the long-term sustainability of the technology. While advancements in hardware, software, and data center optimization are promising, challenges like proprietary model opacity and the rebound effect pose significant hurdles. Companies must prioritize transparency, collaboration, and investment in efficiency technologies to stay competitive and mitigate regulatory risks. The future of AI depends on balancing performance with energy constraints—a challenge that requires innovation at every level.