AI Data Center Power Consumption: The Real Numbers for 2026

Key Numbers
Key Takeaways
- 1Global data centers consumed 415 TWh of electricity in 2024, about 1.5% of total global electricity. IEA projects this reaches 945 TWh by 2030 — more electricity than Japan uses today. The US alone accounts for 183 TWh (4% of US national electricity) in 2024.
- 2AI data center racks draw 60+ kW each, compared to 5-10 kW for standard server racks. This 6-12x density difference is why AI facilities require entirely different power infrastructure, liquid cooling, and grid connections than conventional data centers.
- 3The IEA named AI the most important driver of growth in global data center electricity demand. US capacity is projected to nearly double from 80 GW in 2025 to 150 GW by 2028, requiring construction of new power generation equivalent to dozens of large power plants.
Global data centers consumed 415 terawatt-hours of electricity in 2024. That is roughly 1.5% of all electricity generated on earth. By 2030, the International Energy Agency projects that figure reaches 945 TWh — more than the entire electricity consumption of Japan.
The reason the number is growing so fast is not more data centers. It is what is inside them. A standard server rack draws 5-10 kilowatts. An AI-optimized rack built for NVIDIA H100 or Blackwell GPUs draws 60 kilowatts or more. Put enough of those racks in a building and you need a power connection that used to supply a small city. Meta's Hyperion campus in Louisiana is planned to require 5 gigawatts of continuous power — three times the total electricity consumption of New Orleans.
This article covers how data center power consumption is measured, why AI changed the numbers so dramatically, what the major hyperscalers are actually consuming, and what the grid impact looks like through 2030. The figures come from the IEA, the Lawrence Berkeley National Laboratory, and Bloom Energy's January 2026 infrastructure report.
In This Article
How Data Center Power Consumption Is Measured
Data center power consumption is measured in two ways: total electricity draw in kilowatts (kW) or megawatts (MW) for a single facility, and terawatt-hours (TWh) for industry-wide annual consumption. One terawatt-hour is one trillion watt-hours — the amount of electricity generated by a large power plant running continuously for roughly five months.
The standard efficiency metric is Power Usage Effectiveness (PUE): total facility power divided by IT equipment power. A PUE of 1.0 means every watt entering the building goes directly to servers. A PUE of 1.5 means that for every 100W powering servers, another 50W goes to cooling, lighting, and power distribution overhead. The global average PUE for data centers is approximately 1.58, according to the Uptime Institute 2024 Global Data Center Survey. Hyperscalers do better: Google reports a fleet-wide average PUE of 1.10; Meta has achieved 1.09 at some facilities.
Why PUE Matters Less for AI Facilities
PUE was designed when cooling was the primary overhead cost. In AI data centers, the dynamic shifts. GPUs draw so much power that they generate heat faster than conventional air cooling can remove it. This forces facilities to install liquid cooling systems — direct liquid cooling, immersion cooling, or rear-door heat exchangers — that themselves consume substantial power. A facility with excellent PUE but insufficient cooling capacity simply cannot run AI racks at full load.
The more relevant metric for AI facilities is power density per rack, measured in kW/rack. Standard colocation facilities support 5-10 kW/rack. AI-optimized facilities support 40-80+ kW/rack. This difference drives every infrastructure decision: floor reinforcement, power distribution, cooling capacity, and utility interconnect size. For context on how this affects physical design, the colocation data center explainer covers how facility operators are retrofitting for AI workloads.
| Facility Type | Typical Rack Density | Cooling Method | PUE Target |
|---|---|---|---|
| Standard enterprise | 5-10 kW/rack | Air (CRAC/CRAH) | 1.4-1.6 |
| Hyperscale (pre-AI) | 10-25 kW/rack | Air + hot/cold aisle | 1.1-1.2 |
| AI-optimized | 40-80+ kW/rack | Liquid + air hybrid | 1.05-1.15 |
| Liquid immersion | 80-120+ kW/rack | Full immersion | 1.02-1.05 |
Sources: IEA Energy and AI Report (January 2025), Socomec data center power analysis (2025).
The Global Power Consumption Numbers
The most authoritative source on global data center electricity is the International Energy Agency. Their January 2025 Energy and AI Report, updated April 2025, provides the clearest baseline:
"AI is the most important driver of growth in data center electricity demand and one of the key new energy consumers on a global scale." (International Energy Agency, Energy and AI Report, 2025)
Key figures from the IEA's 2025 Energy and AI Report:
| Year | Global Data Center Consumption | Notes |
|---|---|---|
| 2023 | ~380 TWh | Baseline pre-AI surge |
| 2024 | 415 TWh | 1.5% of global electricity |
| 2026 | 500+ TWh (projected) | TTMS analysis of IEA data |
| 2030 | 945 TWh (projected) | IEA base scenario |
The US numbers, tracked separately by Lawrence Berkeley National Laboratory, show the domestic picture:
- 2018: 76 TWh (1.9% of US electricity)
- 2023: 176 TWh (4.4% of US electricity)
- 2024: 183 TWh (4% of US electricity)
- 2028: 325-580 TWh projected (6.7-12% of US electricity)
That 2028 range is wide because it reflects genuine uncertainty. The lower bound assumes efficiency gains and slower-than-expected AI buildout. The upper bound assumes hyperscaler capacity plans proceed at the announced pace with no offsetting efficiency improvements.
The US capacity picture from Bloom Energy's January 2026 infrastructure report:
- 2025: ~80 GW total US data center demand
- 2028: ~150 GW projected (nearly double in three years)
- 68 hyperscale facilities of 50 MW or larger operational or under construction in the US as of early 2026
Why AI Uses So Much More Power Than Previous Computing
The power difference between AI computing and conventional computing comes down to one thing: parallelism at scale. Training a large language model requires performing billions of matrix multiplications simultaneously across thousands of GPUs. Each GPU is doing useful work every millisecond, drawing full power continuously. Conventional server workloads are highly variable — a web server sits idle most of the time and spikes briefly when requests arrive. An AI training cluster runs at near-100% utilization for weeks.
GPU Power Draw vs Standard Server
A standard 1U server with dual Intel Xeon processors draws approximately 300-500W at peak load. An NVIDIA H100 SXM5 GPU alone draws 700W at maximum thermal design power. A full H100 server (DGX H100) with eight GPUs draws approximately 10,000W (10 kW) from a single server. A rack of 10 DGX H100 servers draws 100 kW — 10-20x a standard server rack.
The newest generation is higher still. NVIDIA GB200 NVL72 rack systems, which package 72 Blackwell GPUs together, draw up to 120 kW per rack unit. Data centers building for GB200 are designing for densities that did not exist in any commercial facility two years ago.
Inference vs Training Power
Training runs are power-intensive but time-bounded. A GPT-4 class training run lasts weeks to months and then ends. Inference — serving model responses to users — runs continuously as long as the product exists. As AI products scale to hundreds of millions of users, inference power demand accumulates indefinitely.
Industry estimates suggest that inference already accounts for 60-70% of AI compute power draw at large AI companies, with training accounting for 30-40%. As model deployment scales faster than new training runs begin, inference will dominate further. The power trajectory is not a bump from training — it is a permanent baseline increase from inference.
The Cooling Problem
Standard air cooling works up to about 30 kW/rack before it becomes impractical. At 60+ kW, air cooling requires so much airflow that it consumes significant power itself and creates hot spots regardless. This is why AI data centers are deploying:
- Direct liquid cooling (DLC): coolant pipes routed directly to GPU heatsinks, removing heat at the source
- Rear-door heat exchangers: liquid-cooled doors that capture exhaust heat before it enters the hot aisle
- Immersion cooling: servers submerged in dielectric fluid
Each cooling method adds infrastructure cost and, in some cases, additional power draw for pumps and chillers.
Hyperscaler Power Consumption and Announced Plans
The five major AI infrastructure spenders — Microsoft, Google, Meta, Amazon, and Oracle — collectively announced over $300 billion in data center capital expenditure for 2025-2026. Power consumption follows capex at roughly proportional scale.
| Company | Known Power / Plans | Source |
|---|---|---|
| Meta | Hyperion campus (Louisiana): 5 GW continuous requirement — 3x total power consumption of New Orleans | Institute on Taxation and Economic Policy, 2025 |
| Microsoft | Targeting 10.5 GW of new data center capacity by 2030; Three Mile Island nuclear plant restarted to supply power | Microsoft investor relations, 2024 |
| 2024 total energy use: 24.3 TWh (up 13% YoY); fleet PUE 1.10 | Google Environmental Report, 2025 | |
| Amazon / AWS | Investing $150B in data center infrastructure through 2028; purchased nuclear power agreements | AWS press releases, 2024-2025 |
| Oracle | Announced 1 GW+ data center campus plans; engaged with nuclear power providers | Oracle investor day, 2025 |
The Number Most Guides Don't Show
Meta's 5 GW Hyperion requirement is often cited as a large number without context. Here is what 5 GW means in practical terms. The US has approximately 900 GW of total installed generating capacity. Meta's single campus requires 0.56% of total US generation capacity running continuously — for one company's one campus.
At $10-12 million per MW for AI-optimized data center construction, 5 GW of capacity represents $50-60 billion in construction cost. That is before equipment. The Hoover Dam generates approximately 2 GW at peak output. Meta's single campus would require the equivalent of 2.5 Hoover Dams running continuously to stay powered.
The grid implications have prompted multiple US states to fast-track new generation approvals specifically for data center demand. Virginia, which hosts the largest concentration of data centers in the world in Northern Virginia, has seen utility applications for new gas peaker plants and solar farms driven primarily by data center contracts.
The Grid Impact of AI Data Center Growth
US electricity demand had been flat or declining for two decades before 2023, as efficiency improvements in lighting, appliances, and manufacturing offset economic growth. AI data center construction has reversed that trend. The US grid was not designed for the scale of new demand now arriving.
The Belfer Center at Harvard Kennedy School published a detailed analysis in 2025 identifying AI data centers as a watershed moment for the US electric grid. The core challenge: building new generation capacity takes 5-10 years. Data center construction takes 18-36 months. The gap creates a period where demand grows faster than supply can be built to meet it.
Consequences already appearing in 2025-2026:
- Power purchase agreements (PPAs) for renewable energy are being signed years in advance, as hyperscalers lock up grid capacity before it is built
- Nuclear power is attracting significant interest from AI companies. Microsoft's deal to restart Three Mile Island's Unit 1 reactor (renamed Constellation Energy's Crane Clean Energy Center) was directly tied to data center power needs, announced September 2023 and operational from late 2024
- Some data center developers are building behind-the-meter generation — natural gas turbines or fuel cells on-site — to avoid grid interconnection queues that now stretch 4-7 years
- States including Georgia, Texas, and Virginia have proposed data center-specific utility rate structures to recover grid upgrade costs from large users
For the broader infrastructure context, the what are AI data centers explainer covers facility design and the AI water usage explainer covers the cooling water dimension of the same infrastructure challenge.
Efficiency Improvements and the 2030 Outlook
The IEA's 945 TWh projection for 2030 is a base case, not a certainty. It assumes current trends in AI adoption, model size, and inference growth continue without major efficiency breakthroughs. Two factors could narrow the gap significantly:
Model Efficiency Improvements
The compute required per unit of AI capability has been falling. DeepSeek R1 demonstrated in January 2025 that comparable reasoning performance to much larger models could be achieved with substantially less compute. If efficiency gains continue at historical rates, the effective AI compute per TWh increases, meaning less energy is needed for the same volume of AI capability.
Hardware Efficiency Gains
NVIDIA's Blackwell generation GPUs deliver approximately 4x better energy efficiency per FLOP compared to the H100 generation for inference workloads. As older H100 infrastructure is replaced by Blackwell and future architectures, the energy per inference request falls. However, the Jevons paradox applies: cheaper, more efficient compute historically leads to more total compute being used, not less.
The IEA Projection Range
The IEA publishes a range, not a single number. Their high scenario reaches approximately 1,100 TWh by 2030; their low scenario reaches around 700 TWh. The 945 TWh base case splits the difference. The range reflects uncertainty about:
- How quickly AI inference scales with model deployment
- Whether efficiency gains in hardware offset demand growth
- Whether regulatory pressure on data center power use materializes in major markets
- The pace of renewable energy and nuclear power deployment specifically for data center supply
What is not in dispute is the direction. US data center electricity consumption grew from 76 TWh in 2018 to 183 TWh in 2024 — a 141% increase in six years. Whatever the exact 2030 number, the trajectory points substantially higher than today.
Frequently Asked Questions
How much electricity do data centers use?
Global data centers consumed 415 TWh of electricity in 2024, approximately 1.5% of total global electricity generation (IEA, 2025). US data centers alone accounted for 183 TWh in 2024, equal to 4% of total US electricity consumption. The IEA projects global data center consumption reaches 945 TWh by 2030, which would exceed the current total electricity consumption of Japan.
Why do AI data centers use so much more power?
AI data centers use more power because AI-optimized GPU racks draw 60+ kW each, compared to 5-10 kW for standard server racks. A single NVIDIA H100 GPU draws 700W at full load; a full DGX H100 server with eight GPUs draws approximately 10 kW. AI workloads also run at near-100% GPU utilization continuously, unlike conventional server workloads that spike briefly and sit idle most of the time. The combination of higher per-rack power and near-continuous utilization multiplies power draw compared to traditional computing.
What is PUE and why does it matter for data centers?
PUE (Power Usage Effectiveness) is total facility power divided by IT equipment power. A PUE of 1.0 means every watt entering the building powers servers. A PUE of 1.5 means 50% overhead goes to cooling and other systems. The global average is approximately 1.58. Major hyperscalers achieve 1.09-1.10. For AI data centers, PUE remains important but power density per rack (kW/rack) has become equally critical, since conventional cooling systems cannot handle 60+ kW racks regardless of PUE target.
How much power does Meta's AI data center use?
Meta's Hyperion campus in Louisiana is planned to require at least 5 gigawatts (GW) of continuous power, according to the Institute on Taxation and Economic Policy (2025). For context, 5 GW is approximately three times the total electricity consumption of New Orleans, and equivalent to the output of 2.5 Hoover Dams running at peak capacity. At $10-12 million per MW construction cost, the facility represents $50-60 billion in infrastructure investment before equipment.
How much will data center power consumption grow by 2030?
The IEA projects global data center electricity consumption will reach 945 TWh by 2030, up from 415 TWh in 2024 — a 128% increase in six years. For the US, Lawrence Berkeley National Laboratory projects consumption of 325-580 TWh by 2028, up from 183 TWh in 2024. Bloom Energy's January 2026 report projects US data center capacity nearly doubling from 80 GW in 2025 to 150 GW by 2028. All projections reflect meaningful uncertainty depending on AI adoption pace and hardware efficiency improvements.
Are AI companies switching to renewable or nuclear power?
Yes. Microsoft restarted the Three Mile Island nuclear reactor (renamed Crane Clean Energy Center) in late 2024, with output contractually dedicated to data center power. Amazon has signed multiple nuclear power purchase agreements and is investing in small modular reactor development. Google signed agreements for geothermal and nuclear power. Meta has announced renewable energy contracts totaling multiple GW. Many AI companies are also building behind-the-meter natural gas generation on-site to bypass grid interconnection queues that now stretch 4-7 years.
What percentage of data center power goes to GPUs vs cooling?
In AI data centers, approximately 60% of total electricity goes to IT equipment (primarily GPUs and servers), with the remaining 40% covering cooling systems, power distribution, lighting, and other overhead. This ratio improves as PUE improves. At a PUE of 1.10 (Google's fleet average), 91% of power reaches IT equipment and only 9% is overhead. At a PUE of 1.5 (industry average), 67% reaches IT equipment and 33% is overhead. Within the IT equipment share, GPUs in AI-optimized facilities account for 70-85% of server power draw.