Skip to main content
Insights8 min read

The $690B Moat: Why Hyperscaler Capex Is the Best Argument for Local AI

February 24, 2026 by Asif Waliuddin

AI
The $690B Moat: Why Hyperscaler Capex Is the Best Argument for Local AI

The $690B Moat: Why Hyperscaler Capex Is the Best Argument for Local AI

Microsoft, Google, Amazon, Meta, and Oracle are collectively spending $660-690 billion on AI infrastructure in 2026. Nearly double what they spent in 2025. Over 75% of it earmarked for AI chips, servers, and data centers. Google alone disclosed a capex range of $175-185 billion -- up 97% year-over-year from $91.4 billion. Microsoft is committing $120 billion or more. The Stargate project has $400 billion+ committed through 2029.

These are the largest capital expenditures in the history of enterprise technology. They are being reported as evidence that AI demand is real and the hyperscalers are positioned to capture it.

I want to offer a different reading of the same numbers. One that the hyperscalers would prefer you not think about too hard.

The Hype

The $690 billion narrative is triumphal. AI demand is insatiable. The hyperscalers are responding with unprecedented investment. The infrastructure is being built. The future is being poured in concrete and wired with fiber. Every capex chart published in the last quarter carries the implicit message: this is what confidence looks like.

Analysts cite the spend as validation. Executives cite it as strategic commitment. Press releases frame each new data center announcement as proof that AI is not a bubble -- it is a buildout. The implication is clear: the companies that can spend $690 billion are the companies that will own the AI infrastructure layer. And owning the AI infrastructure layer means owning the AI economy.

That framing is not wrong. It is incomplete. And what it leaves out is the part that matters most to developers.

The Reality

The $690 billion is not building infrastructure for everyone. It is building infrastructure for the customers who can afford to access it on hyperscaler terms.

Here is the math. When five companies spend $690 billion on AI infrastructure, that capital must generate returns. Returns come from customers. The customers who generate the largest returns -- Fortune 500 enterprises signing multi-year, multi-hundred-million-dollar contracts -- get priority access to capacity, to support, to new capabilities. This is not speculation. It is how capacity-constrained businesses operate. Microsoft has $80 billion in unfilled Azure orders right now, and the allocation logic is straightforward: serve the largest contracts first.

For the developer who is not enterprise-scale, the $690 billion buildout creates a paradox. More AI infrastructure is being built than at any point in history, and yet accessing that infrastructure on terms that make sense for a startup, an indie developer, or a mid-market company becomes harder, not easier, as the capital concentration deepens.

Three specific mechanisms:

Pricing follows concentration. When five companies control the AI infrastructure layer and demand exceeds supply, pricing power shifts entirely to the vendor. The current price of a GPU-hour on Azure or GCP is not set by competition. It is set by scarcity. As long as the $80B-scale backlogs persist, there is no incentive to compete on price. There is every incentive to price to the constraint. Developers without enterprise procurement leverage pay the most per unit of compute.

Access follows contract size. Capacity-constrained hyperscalers prioritize their largest customers. This is operationally rational and economically obvious. But it means that the developer who needs 100 GPU-hours next month for a product launch is competing for capacity against the enterprise that reserved 10 million GPU-hours last year. The queue is invisible, but it is real. Microsoft's $80 billion backlog is the queue made visible.

Innovation follows the money. When AI infrastructure is concentrated in five companies spending $690 billion, the research, tooling, and platform capabilities follow. The hyperscalers build for their highest-value customers. Enterprise compliance tooling. Fortune 500 security frameworks. Multi-year contract SLA guarantees. The developer tooling that serves individual builders and small teams is an afterthought -- or, more precisely, a go-to-market motion designed to create future enterprise customers. The free tier is not generosity. It is a pipeline.

The Moat

The $690 billion is not just capital deployment. It is moat construction.

Every new data center, every exclusive chip contract with NVIDIA or AMD, every long-term power purchase agreement -- these are structural barriers that make it progressively harder for anyone outside the hyperscaler tier to offer comparable AI infrastructure. The moat is not protecting users. It is protecting the infrastructure owners.

And the moat is getting wider. Google's capex increased 97% in a single year. Microsoft is spending $120 billion in one year. The Stargate project envisions 5GW+ data center clusters under US national security coordination. These are not just investments in infrastructure. They are investments in making alternative infrastructure irrelevant.

Governments have noticed. The EU, Saudi Arabia, India -- sovereign AI programs are launching specifically because these nations recognized that depending on hyperscaler infrastructure means depending on the commercial and geopolitical priorities of the companies that own it. When nation-states are building sovereign compute because they do not trust the concentration of AI infrastructure in five American companies, individual developers should ask the same question: is depending on this concentrated infrastructure in my interest?

What Local-First AI Actually Means in This Context

Local-first AI is not a lifestyle choice. It is the structural alternative to the $690 billion moat.

When AI runs on hardware you control -- your machines, your GPUs, your infrastructure -- three things change:

The pricing is yours. You pay for hardware once and run inference indefinitely. There is no per-token cost that scales with usage. There is no vendor who can reprice your access because demand exceeds their capacity. The economics of local AI are CAPEX-forward, not OPEX-dependent. For any workload that runs repeatedly, the cost math favors local infrastructure within months.

The access is immediate. There is no queue. There is no allocation lottery. There is no dependency on whether Microsoft can bring a data center online in time for your product launch. Your deployment timeline is governed by your readiness, not a vendor's infrastructure buildout schedule. In a world where the largest cloud vendor has $80 billion in unfilled orders, "it runs when you tell it to run" is not a feature. It is a competitive advantage.

The stack is portable. Models that run locally are not vendor-locked. Open-weight models -- Llama, Mistral, Qwen, DeepSeek, and the dozens of capable alternatives released every month -- run on commodity hardware. The model layer is commoditizing. BlackRock's AI-focused ETF is rotating away from model vendors toward infrastructure specifically because institutional capital sees this commoditization happening. The durable value is not in any specific model. It is in the infrastructure and tooling layer that makes commodity models useful.

The $690B Argument

Here is the part the hyperscalers do not want you to hear: the bigger their capex, the stronger the case for local-first AI.

At $100 billion in combined hyperscaler spend, you could argue that centralized AI infrastructure was the efficient default. Economies of scale. Shared infrastructure. Elastic pricing. The cloud pitch worked when the infrastructure was abundant relative to demand.

At $690 billion, the infrastructure is no longer abundant. It is constrained, allocated, queued, and priced to the constraint. The economics that made cloud AI attractive at smaller scale invert at $690 billion scale. The capital concentration creates the scarcity. The scarcity creates the pricing power. The pricing power creates the access barrier. And the access barrier is the argument for building on infrastructure you control.

$690 billion is not evidence that cloud AI is winning. It is evidence that cloud AI is becoming a gated resource -- and the gate is getting more expensive to pass through every quarter.

The Bottom Line

Microsoft spent $120 billion this year to solve a delivery backlog it created by selling AI infrastructure faster than physics allows it to build. Google increased capex by 97% to stay in a race where the finish line is whoever builds the most data centers fastest. The Stargate project is building 5GW+ data center clusters under national security coordination.

None of this infrastructure is being built for you. It is being built for the enterprise customers who signed the contracts that justify the capital. If you are not one of those customers, you are in line behind them. And the line is $80 billion long.

Local-first AI exists because the $690 billion buildout does not reach most developers. That is not a criticism. It is a market structure observation. And it is the reason that AI infrastructure you own -- that runs on hardware you already have, deploys on your timeline, and does not require you to bid against sovereign governments for GPU access -- is not a niche preference.

It is the rational response to a $690 billion moat.