Data Center Alley almost went dark in March 2025.
A cluster of AI data centers near Fairfax, Virginia (the ones serving Microsoft, Google, and Amazon) created such wild power fluctuations that the electrical grid nearly collapsed. Grid operators caught it just in time.
Here's what that near-miss tells us:
Every time you use ChatGPT, you're pulling on a supply chain that's quietly breaking. From power plants to chip fabs to cooling systems, every link is straining.
And it's getting worse.
Let me walk you through what's actually happening.
The power problem is worse than you think
A single hyperscale AI data center needs 100 to 500 megawatts. That's a medium-sized city's worth of power.
Here's the mismatch: Grid planning takes 5-10 years. Permitting. Environmental reviews. Construction. The whole bureaucratic dance. AI companies want power now.
You can build the data center. You can buy the GPUs. But if the grid can't deliver electrons, you have expensive hardware sitting idle.
By 2030, data centers will eat 945 terawatt-hours globally (more electricity than Japan consumes in a year). AI workloads are 10% of that today. They'll hit 20% by 2030.
But here's what makes this tricky: The power demand changes completely depending on what you're doing with AI.
Training vs inference: two totally different beasts
AI infrastructure is really two separate supply chains with opposite needs.
Training = massive compute explosions
When OpenAI trains GPT-5, thousands of H100 GPUs run full blast for months. Think of it like running a small city at maximum capacity for weeks on end.
What that needs:
The newest, fastest GPUs money can buy
Thousands of chips working together
Industrial-scale cooling (the heat is insane)
High-speed links between every chip
Massive upfront spending ($400B in 2025, projected $550B in 2026)
Inference = steady hum of usage
When you ask ChatGPT a question, that's inference (running an already-trained model). Unlike training's explosive bursts, inference is predictable. Millions of users querying all day creates a steady load.
What that needs:
Older or specialized chips work fine
Servers spread across many locations
Fast connections to users
Ongoing costs that never stop
Data centers love inference because they can plan for it. Training terrifies grid operators because the load spikes unpredictably.
Five layers from power plant to your answer
Let me trace the whole chain. Each layer has its own breaking point.
Layer 1: Electricity
It starts with power generation. Data centers used 1.5% of global electricity in 2024, growing 12% annually for five years straight.
That growth can't continue without massive grid upgrades.
Here's the scary part: Only 60% of projected 2030 demand has identified power sources. The other 40%? Nobody knows where it's coming from.
Layer 2: Data centers
The power flows into facilities that need:
Access to massive electricity capacity
Industrial cooling systems (GPUs generate concentrated heat)
Physical security and redundancy
High-speed network connections
Data Center Alley in Virginia became a hub because of grid access. Now that access is maxed out. New facilities are moving to Texas, Arizona or anywhere with spare power capacity.
Layer 3: GPUs and specialized chips
NVIDIA owns 94% of the AI GPU market. Their H100, H200, and new Blackwell chips run most AI training.
The manufacturing is a global relay race:
TSMC makes chips in advanced fabs (now including Phoenix, Arizona)
Chips ship to Taiwan for packaging
Packaged chips go to assembly facilities
Assembled servers finally reach data centers
Even Blackwell chips made in Arizona ship to Taiwan for packaging, then back to Texas for assembly. The geopolitical risk is obvious.
AMD is emerging as competition with multi-billion dollar OpenAI deals. Broadcom is building custom inference chips. But NVIDIA's CUDA software creates massive lock-in: switching means rewriting all your code.
Layer 4: Training data
AI models eat massive datasets: text, images, videos from everywhere. This data goes through:
Collection (web scraping, licenses, proprietary sources)
Cleaning and preprocessing (takes serious compute)
Storage with high-speed access
Continuous updates
We're talking petabyte-scale data pipelines. Storage infrastructure needs both redundancy and instant access.
Layer 5: Foundation models
Modern AI models have billions of parameters. Training one requires:
Weeks or months on thousands of GPUs
Orchestrating distributed systems
Iteration and testing
Once trained, the model runs inference: generating your responses. Inference uses way less compute per query but runs billions of times. It optimizes for speed (you expect instant responses) and increasingly runs on edge locations closer to users.
Training centralizes in massive GPU clusters. Inference distributes everywhere.
Three bottlenecks choking the supply chain
Power grid capacity is the hard wall
You can make more GPUs. But if the grid can't deliver power, they're paperweights.
Grid expansion needs 5-10 years. Transmission lines. Substations. Permits. Coordination across utilities and regulators.
AI development moves in months. That's the crisis.
Chip manufacturing is concentrated in Taiwan
TSMC controls advanced chip production. The geopolitical risk is obvious. New Arizona fabs won't hit full capacity until 2026-2027.
Each fab needs:
Billions in capital
Years to reach volume production
Rare equipment (ASML's lithography machines)
Complex supply chains for materials
Even with new capacity, demand outpaces supply.
Specialized engineering talent is scarce
You can't hire your way out of this fast. Data center design and cutting-edge chip manufacturing require expertise that takes years to develop. Talent constraints limit how quickly new facilities can scale.
What this means if you're building in AI
The supply chain reality creates hard choices:
Companies scramble for power commitments before breaking ground. Data center location depends more on electricity availability than network speed or real estate. NVIDIA dominates not just because their chips are faster, but because CUDA lock-in makes switching painful.
The infrastructure buildout is approaching half a trillion dollars in 2026. But it's slamming into physical limits on power and manufacturing.
Here's what's happening every time you use AI
When you ask ChatGPT a question, you're tapping into this massive, strained system.
Your answer represents: Electrons from power plants → transmission lines → data centers → GPUs manufactured across continents → data collected globally → models trained on thousands of chips → response in milliseconds.
Every layer is straining. Grids can't expand fast enough. Chip manufacturing clusters in geopolitically risky locations. Engineering talent stays scarce.
That Data Center Alley near-miss in March? Not a one-off. It's what happens when AI infrastructure pushes electrical systems to their limit.
Every AI response you get comes from a supply chain operating at the edge of what's possible. The electrons answering your questions are getting more expensive and harder to find.
That's going to shape what AI can do next.