Who Pays? The $220B Capex Explosion
Post 7: Terrestrial Foundation (SECTION 1 FINALE)
Where Hyperscalers Spend — And When the Music Might Stop
By Randy Gipe | March 2026
But somebody has to pay for all of it.
Microsoft, Google, Amazon, Meta: $220 billion in combined capex for 2025. That’s $600 million per day. Every single day.
And it’s all flowing to AI infrastructure.
This is the final piece of the terrestrial foundation: Who’s funding the boom—and what happens if they stop?
Part 1: The Big Four — $220B in 2025
Hyperscaler Capex Breakdown
| Company | 2024 Capex | 2025 Capex (est.) | AI Share | Primary Use |
|---|---|---|---|---|
| Microsoft | ~$55B | $65-70B | ~70% | Azure AI, OpenAI partnership |
| Google (Alphabet) | ~$50B | $60-65B | ~65% | Gemini, Cloud AI, TPUs |
| Amazon (AWS) | ~$55B | $60-65B | ~60% | AWS AI services, Trainium |
| Meta | ~$30B | $35-40B | ~75% | Llama models, AI infra |
| TOTAL | ~$190B | ~$220-240B | ~65-70% | AI dominates |
$220B = More than the GDP of New Zealand.
Where it goes (average allocation):
- GPUs + servers: 40-50% (~$88-110B)
- Networking: 20-30% (~$44-66B)
- Power + cooling: 15-20% (~$33-44B)
- Buildings + land: 10-15% (~$22-33B)
Microsoft — The OpenAI Bet
💻 MICROSOFT: $65-70B CAPEX (2025)
Why so high:
- OpenAI partnership: Exclusive cloud provider for ChatGPT/GPT-4/GPT-5
- Microsoft funds OpenAI's compute via Azure credits
- Azure AI growing 30%+ YoY (copilots, enterprise AI)
Where it goes:
- NVIDIA GPUs: 100,000+ H100/H200/Blackwell (estimated)
- Data centers: Building 50-100 new facilities globally (2024-2026)
- Nuclear power: $16B Three Mile Island restart (see Post 8)
Revenue from AI (2025):
- Azure AI revenue: ~$10-15B (growing, but not yet covering capex)
- Microsoft 365 Copilot: $30/user/month (millions of users, ramping)
The ROI question:
- Spending $65-70B/year
- AI revenue: ~$15-25B
- Not profitable yet, but betting on future growth
Google — Defending Search
🔍 GOOGLE: $60-65B CAPEX (2025)
Why spending:
- Existential threat: ChatGPT potentially disrupts Google Search
- Gemini models competing with ChatGPT/Claude
- Google Cloud AI services (enterprise customers)
Strategy:
- Mix of NVIDIA GPUs + custom TPUs (diversified, less NVIDIA-dependent)
- Building data centers globally (U.S., Europe, Asia)
- TPU v5p optimized for Gemini training
Revenue from AI:
- AI-enhanced search ads (incremental, hard to isolate)
- Google Cloud AI: $5-10B (growing 40%+ YoY)
Advantage:
- Search still prints $200B+/year in advertising → can fund AI indefinitely
- Not dependent on AI profitability short-term
Amazon — AWS Dominance
☁️ AMAZON: $60-65B CAPEX (2025)
Why spending:
- AWS = cloud leader (32% market share)
- Enterprise customers demanding AI services
- Competing with Azure AI, Google Cloud
Strategy:
- NVIDIA GPUs for customer workloads
- Custom Trainium/Inferentia chips (cost advantage for inference)
- 1.9 GW nuclear power (Susquehanna PPA, see Post 8)
Revenue from AI:
- AWS AI services: ~$10-20B (growing 50%+ YoY)
- Bedrock (foundation model API): Ramping
Advantage:
- AWS already profitable ($90B+ revenue, $30B+ operating income)
- AI capex funded by existing cash cow
Meta — Open Source Llama
📘 META: $35-40B CAPEX (2025)
Why spending:
- Llama models (open source, but Meta trains them)
- AI for Facebook/Instagram feeds (recommendations, ads)
- Metaverse pivot failed, AI is new priority
Strategy:
- 350,000+ H100 GPUs (announced goal by end 2024, expanding)
- Building own data centers (not leasing)
- 6.6 GW nuclear power RFPs (see Post 8)
Revenue from AI:
- No direct AI product sales (Llama is free)
- AI improves ad targeting → incrementally higher ad revenue (~$150B+ total)
The risk:
- Highest AI capex as % of revenue (no separate AI revenue stream)
- Betting AI improves core ads business enough to justify spend
Part 2: The AI Startups — Burning Cash on Compute
OpenAI — The $6B Annual Burn
OpenAI revenue (2025 est.): $3-4B
- ChatGPT subscriptions: $20/month × millions of users
- Enterprise API usage
OpenAI costs (2025 est.): $9-10B
- Compute (Azure credits from Microsoft): ~$6-7B
- Salaries, R&D, operations: ~$3B
Annual burn: ~$6B
How it's funded:
- Microsoft Azure credits (part of partnership)
- Equity raises ($10B+ from Microsoft, others)
- Revenue doesn't cover costs yet
Path to profitability:
- Need $10B+ revenue (3x current)
- Or reduce compute costs via efficiency/cheaper chips
- Timeline: 2027-2028 (if growth continues)
Anthropic — $3-4B Burn
Anthropic revenue (2025 est.): $1-2B
- Claude subscriptions + API
- Enterprise deals
Anthropic costs (2025 est.): $5-6B
- Compute: ~$3-4B (AWS + Google Cloud)
- Salaries, R&D: ~$2B
Annual burn: ~$3-4B
Funding:
- $7.3B raised (Google, Amazon, others)
- Runway: 2-3 years at current burn
xAI, Cohere, Inflection, Others
Collective burn: $5-10B/year
- xAI (Musk): $10B raise, building 100k GPU cluster in Memphis
- Cohere, Inflection, Character.AI, others burning $500M-2B each
Total AI startup burn (2025): $15-20B/year
None are profitable yet.
Part 3: The ROI Question — When Do Returns Materialize?
Current State (2025-2026)
⚠️ AI REVENUE vs. CAPEX GAP
Total AI infrastructure spending (2025):
- Hyperscalers: $220B capex
- Startups: $15-20B burn
- Total: ~$240B/year
Total AI revenue (2025 est.):
- Cloud AI services (Azure, AWS, GCP): $25-45B
- AI app subscriptions (ChatGPT, Claude, etc.): $5-10B
- Enterprise AI software: $10-20B
- Total: ~$40-75B
Gap: Spending $240B, earning $40-75B → $165-200B deficit
This is fine IF revenue grows fast enough to catch up.
But if it doesn't...
Bull Case — Revenue Catches Up (2027-2030)
Scenario: AI becomes as transformative as cloud computing.
Cloud revenue trajectory (2010-2020):
- Early years: Massive capex, minimal revenue
- 2015+: Revenue inflection, capex still high but profitable
- 2020: AWS $45B revenue, $13B profit
AI could follow same path:
- 2025-2026: Capex > revenue (current state)
- 2027-2028: Revenue inflection (enterprises adopt AI at scale)
- 2029-2030: AI revenue $150-250B, profitable
What needs to happen:
- ChatGPT/Claude usage grows 5-10x (more paying users)
- Enterprise AI adoption accelerates (Microsoft Copilot in every company)
- New use cases emerge (AI agents, autonomous workflows)
Bear Case — Revenue Stalls (2027-2028 Capex Taper)
Scenario: AI hits plateau, revenue doesn't justify capex.
Warning signs:
- ChatGPT growth slowing (user saturation)
- Enterprises skeptical of AI ROI (hype > reality)
- Hyperscalers cut capex 10-20% (2027-2028)
What happens:
- NVIDIA revenue drops 20-30% (hyperscalers main customers)
- Data center REITs see lease slowdown
- Vertiv, Arista, Broadcom all impacted
- AI startups run out of runway, consolidate or shut down
Historical precedent:
- Dot-com bubble (2000): Massive capex, revenue didn't materialize, crash
- Crypto mining (2018): Capex boom, then crash, stranded infrastructure
Probability: 20-30% chance of significant taper by 2028
Part 4: Section 1 Synthesis — The Complete Terrestrial Stack
🏗️ COMPLETE TERRESTRIAL FOUNDATION (Posts 1-7)
Post 1: NVIDIA
- $130B+ revenue, 75% margins, 80%+ market share
- CUDA moat = 18-year lock-in
- Blackwell 2x performance but 30% more power
Post 2: TSMC
- Only company making 5nm/3nm at scale
- NVIDIA 100% dependent (no backup plan)
- Arizona 70% yields vs. Taiwan 95% (geopolitical risk)
Post 3: Power Crisis
- 945 TWh by 2030 (2.3x growth), 8.9% of U.S. electricity
- 134 GW capacity needed (grids maxing out)
- Consumer bills up 8-25%, political backlash brewing
Post 4: Data Center REITs
- Digital Realty, Equinix: $1B+ leases, 60-70% margins
- Bitcoin miners pivot: IREN $3.4B ARR, CIFR $9.3B contracts
- Power infrastructure = competitive advantage
Post 5: Networking
- 20-30% of AI cluster cost (invisible but critical)
- NVIDIA InfiniBand dominates training (70-80%)
- Arista +150-190%, NVIDIA networking $20-25B revenue
Post 6: Cooling
- Liquid cooling 50% adoption (Blackwell requires it)
- Vertiv +800-1,000%, Schneider 20-30% YoY growth
- 15-20% of data center capex
Post 7: Who Pays
- Hyperscalers: $220B capex (2025)
- AI startups: $15-20B burn
- Revenue gap: $240B spending, $40-75B revenue
- ROI risk: 20-30% chance of taper if revenue stalls
The picks-and-shovels thesis:
- Winners NOW: NVIDIA, TSMC, REITs, Vertiv, Arista (all printing money)
- Losers NOW: AI apps burning cash (OpenAI $6B/year)
- Risk 2027-2028: If AI revenue doesn't catch up, entire infrastructure capex tapers
What's Next in the Series
SECTION 1 COMPLETE: Terrestrial Foundation ✅
SECTION 2 BEGINS: The Power Solution (Posts 8-9)
Post 8 (next): SMR Nuclear Renaissance — Hyperscalers Go Atomic
The power crisis (Post 3) needs a solution. Enter Small Modular Reactors:
What we'll cover:
- Microsoft $16B Three Mile Island restart (835 MW by 2028)
- Google 500 MW Kairos Power SMRs
- Amazon 1.9 GW Susquehanna PPA
- Meta 6.6 GW nuclear RFPs
- Why SMRs = 3-5 year timeline (vs. 10-15 for traditional nuclear)
- 10 GW pipeline by 2030 (20-30% of U.S. data center power)
Then Post 9: Grid Constraints & Utility Scramble
Then Section 3: The Global Race (China, Singapore, Geopolitics)
SOURCES
Hyperscaler Capex:
- Microsoft, Google, Amazon, Meta quarterly earnings (Q4 2025): Capex disclosed in 10-Qs, earnings calls
AI Startup Burns:
- OpenAI, Anthropic: Industry estimates (The Information, Bloomberg reports), funding announcements
Revenue Estimates:
- Azure AI, AWS AI, Google Cloud: Segment revenue from earnings (where disclosed)
- AI app subscriptions: Public user numbers × pricing

No comments:
Post a Comment