Is there a silver bullet to help data centers cope with AI’s energy demands?
There’s no denying the impact AI is already having on data centers and energy use. If left unchecked, the situation is only going to get worse. A recent IDC report reveals that with AI adoption skyrocketing, the energy needed to support AI workloads is set to soar, with data center electricity consumption expected to more than double between 2023 and 2028. AI-driven workloads alone are projected to grow at a staggering compound annual growth rate (CAGR) of 44.7% through 2027, with energy demands reaching a massive 146.2 TWh. The implications are stark – data centers, which already account for 46% of enterprise energy spend, could soon become unsustainable.
That, of course, cannot happen. But with AI workloads rapidly increasing, data centers have to evolve quickly and manage the prospect of a new energy crisis, with rising electricity prices driven by geopolitical instability in the Middle East. The growing influence of AI tools across industries – from healthcare to financial services – is undeniable. However, an AI-powered search uses 100 times more energy than a non-AI-powered search, while building foundational AI models can consume enough energy to power 20,000 homes for six months.
Director of Systems Engineering at Nutanix.
A solution?
An Atlantic Ventures report, Improving Sustainability in Data Centers 2024, suggests a solution, revealing how next-generation data center architectures, like hyperconverged infrastructure (HCI), can reduce energy consumption, lower carbon emissions, and drive cost savings across the EMEA region. In just six years, the report finds that modernizing data centers with HCI could save up to 19 million tCO2e in the EMEA region, equivalent to the emissions of almost 4.1 million cars. It could also save €25 billion by 2030 from improved energy and operational efficiencies.
As organizations integrate AI across operations and come to terms with the sheer scale of energy consumption, HCI could reduce the risk of spiraling costs and ensure sustainability goals are not missed. But it’s not just about HCI, it’s about how organizations work with AI. Focus should shift to optimising where and how AI workloads are processed, using modernisation to manage workloads more intelligently. This makes so much more sense than to just keep building more energy-efficient data centers.
This is important, as we have to take into account how AI works and where the demands for power are going to increase. While many organizations are captivated by the energy consumption required to train foundational AI models, for example, it is inferencing – the real-time decision-making AI performs – where the bulk of the energy is spent.
Foundational model training happens once but inferencing is a continuous process that happens millions of times, especially with AI-driven applications like fraud detection or predictive maintenance. Optimizing inferencing, particularly at the edge, could be the silver bullet data centers need to manage AI energy demands more efficiently.
Turning to renewable energy
As the IDC report suggests, more data center providers need to turn to renewable energy sources, but they also need to rethink their infrastructure. Hybrid cloud, edge computing, and on-premise systems offer a way to balance AI’s energy demands by distributing workloads more intelligently.
Sign up to the TechRadar Pro newsletter to get all the top news, opinion, features and guidance your business needs to succeed!
Processing data closer to its source with edge computing, for example, reduces the energy needed to transfer large datasets back and forth from centralized servers. Meanwhile, hybrid cloud computing environments can handle the computationally intense AI training tasks, leaving real-time inferencing to occur on-premise or at the edge.
Edge computing also plays a pivotal role by processing data closer to where it’s generated, such as in retail stores or IoT devices. This not only improves response times but also significantly reduces the energy required for inferencing.
Modern infrastructure is crucial to managing AI’s energy demands and a containerized platform, designed to handle both CPUs and GPUs, is necessary to run AI workloads efficiently. Storage also becomes critical, as AI typically deals with unstructured data like files and objects. By investing in high-performance storage systems and optimized compute stacks, businesses can significantly reduce the energy required to run AI applications.
Moreover, the ability to measure and manage energy consumption is critical. Platforms that provide real-time visibility into energy use enable data centers to optimize every stage of AI processing – from training to inferencing. Even a 10% improvement in energy efficiency, according to the IDC report, can lead to substantial savings.
Real-time decision-making
Rather than focusing solely on the immense energy costs of training foundational models, businesses need to pay attention to how often these models are used in real-time decision-making. Compressing models, refining their structure, and running them on platforms designed for efficiency will be key in reducing AI’s overall energy footprint. For example, we have developed container platforms and high-performance storage solutions specifically tailored for AI inferencing, offering businesses a way to optimize their AI workloads and temper their energy demands.
The true cost of AI is no longer just about performance and innovation, it’s about the energy required to sustain it. As organisations ramp up their AI initiatives, the question isn’t whether they can afford to invest in AI but whether they can afford the energy it consumes. With hybrid infrastructure and a focus on efficient inferencing, businesses have a way to temper this energy surge. Otherwise, those that ignore this reality may soon find their data centers at the mercy of an AI energy crisis.
We’ve featured the best colocation providers.
This article was produced as part of TechRadarPro’s Expert Insights channel where we feature the best and brightest minds in the technology industry today. The views expressed here are those of the author and are not necessarily those of TechRadarPro or Future plc. If you are interested in contributing find out more here: https://www.techradar.com/news/submit-your-story-to-techradar-pro
There’s no denying the impact AI is already having on data centers and energy use. If left unchecked, the situation is only going to get worse. A recent IDC report reveals that with AI adoption skyrocketing, the energy needed to support AI workloads is set to soar, with data center…
Recent Posts
- Meta turns to nuclear energy for its AI ambitions
- 17 Gifts for People Who Are Perpetually Cold (2024)
- Google’s generative AI video model is available in private preview
- Top vodka brand Stoli files for bankruptcy following ransomware attack
- GM and EVgo now have over 2,000 EV charging stalls and counting
Archives
- December 2024
- November 2024
- October 2024
- September 2024
- August 2024
- July 2024
- June 2024
- May 2024
- April 2024
- March 2024
- February 2024
- January 2024
- December 2023
- November 2023
- October 2023
- September 2023
- August 2023
- July 2023
- June 2023
- May 2023
- April 2023
- March 2023
- February 2023
- January 2023
- December 2022
- November 2022
- October 2022
- September 2022
- August 2022
- July 2022
- June 2022
- May 2022
- April 2022
- March 2022
- February 2022
- January 2022
- December 2021
- November 2021
- October 2021
- September 2021
- August 2021
- July 2021
- June 2021
- May 2021
- April 2021
- March 2021
- February 2021
- January 2021
- December 2020
- November 2020
- October 2020
- September 2020
- August 2020
- July 2020
- June 2020
- May 2020
- April 2020
- March 2020
- February 2020
- January 2020
- December 2019
- November 2019
- September 2018
- October 2017
- December 2011