fbpx

Nvidia’s H100 AI GPUs (graphics processing units) are taking the tech world by storm, but their reign comes at the price of a hefty energy bill. According to Stocklytics.com, these power-hungry processors are projected to consume some 13 797 GWh in 2024, surpassing the annual energy consumption of entire nations like Georgia and Costa Rica.

Energy hungry data processing microchips

The soaring energy consumption of AI GPUs raises concerns about the environmental impact and sustainability of AI advancement.
[Photo credit: american-public-power-association-TF-DL_2L1JM-unsplash] 

These findings bring up concerns about the environmental impact and sustainability of AI advancement.

Stocklytics Financial Analyst Edith Reads commented on the analysis: “AI, which often requires running computations on gigabytes of data, needs enormous computing power compared with ordinary workloads. And Nvidia’s cutting-edge H100 AI GPUs are leading the way in escalating energy consumption. Each H100 GPU, running at 61% annual utilisation, consumes roughly 3 740 kilowatt hours (kWh) of electricity annually. This is equivalent to the average American household. While this figure might seem alarming, GPU efficiency may improve in the near future, offering a potential path towards more sustainable computing.”

Venturing into the $30 billion tailored chip market

As a leading player in AI chip design, Nvidia is broadening its scope by venturing into custom chip development for cloud computing and AI applications. The firm is now looking to tap into the growing custom chip sector, projected to reach $10 billion this year and double by 2025. The broader custom chip market hit around $30 billion in 2023, accounting for 5% of chip sales annually.

Based in Santa Clara, California, Nvidia targets the changing needs of tech giants such as OpenAI, Microsoft, Alphabet, and Meta Platforms. The company is establishing a division focused on developing custom chips, including powerful artificial intelligence (AI) processors, for cloud computing firms and other enterprises.

Currently holding 80% of the high-end AI chip market share, the position has driven Nvidia’s stock market value up by 40% so far this year to $1.7 trillion after a more than threefold increase in 2023.

The energy challenge of powering AI chips

As Nvidia’s aspirations grow, concerns are emerging regarding the impact of the escalating energy requirements linked to its cutting-edge chip technologies.

According to Paul Churnock, Microsoft’s Principal Electrical Engineer of Datacentre Technical Governance and Strategy, the installation of millions of Nvidia H100 GPUs will consume more energy than all households in Phoenix, Arizona by the end of 2024.

Successfully navigating these challenges and fostering innovation will shape the future landscape of AI computing and beyond. Amazon’s recent unveiling of the Arm-based Graviton4 and Trainium2 chips holds promise for efficiency gains.

For more information visit: https://stocklytics.com