The Operating Manual for the AI Revolution
The biggest industrial shift in modern history is already underway, and if you cannot see the full stack, you cannot see where the world is heading.
How $60B of AI revenue generates $700B of infrastructure investment
72 Rubin servers + 256 LPU chips = 700M tokens/sec — 350× Hopper throughput. Purpose-built for inference.
Blackwell (now) → Vera Rubin (2026, HBM4) → Feynman (2028, TSMC 1.6nm, silicon photonics)
The application layer is the fastest-growing segment of AI revenue.
Historically, in every major computing wave — mainframes, PCs, mobile, cloud — the application layer is where most value ultimately accrues. Infrastructure enables, but applications capture.
Source: company disclosures, press reports, DVC analysis. “Pure application layer” = companies with own distribution, excluding model providers and generative media.
Time for each software category to reach ~$60B in combined application revenue
Source: Bessemer Cloud Index, Menlo Ventures, DVC analysisTop AI application companies by annualized recurring revenue
API-heavy = Most revenue from API/platform | End-user = Most revenue from consumers/enterprise seats
Capital is flooding every layer, but one battleground shapes the rest. To understand power in AI, you have to understand the fight over the models themselves.
10+ companies at the frontier simultaneously. The bottom is commoditizing; the top is getting more expensive and more strategic. Open-source models now match proprietary ones on most benchmarks.
GPT-5.4 / o3
$25B ARR
Opus 4.6 / Mythos
#1 Chatbot Arena
Gemini 3.1 Pro
#2 Chatbot Arena
Grok 3
$1.25T merged
Nova 2 Pro
Top reasoning on Bedrock
Muse Spark
Closed · 3.6B DAU
Llama 4 Maverick
Community License
V3.2 / R1
Open Weight · MIT
Qwen 3.5 397B
Apache 2.0
GLM-5 744B
#1 OS leaderboard
Kimi K2.5 1T
99.0 HumanEval
Mistral Large 2
EU Sovereign AI
Nemotron 3
Open Agentic
Gemma 4
Apache 2.0
GPT-oss 120B
Apache 2.0
Step-3.5-Flash
97.3 AIME
M2.5 230B
80.2 SWE-bench
THE FOUR SCALING LAWS OF AI — Jensen Huang, Lex Fridman Podcast #494 (Mar 2026)
Bigger models + more data + more compute = smarter AI. The original scaling law.
Synthetic data, RLHF, fine-tuning, distillation. “We are no longer limited by data — we are limited by compute.”
100x+ compute at inference for multi-step reasoning. “Inference is thinking, and thinking is hard.”
Agents spawn sub-agents, use tools, create data. “It’s like multiplying AI. We could spin off agents as fast as you want.”
“Intelligence is going to scale by one thing, and that’s compute.”
Source: Lex Fridman Podcast #494, NVIDIA GTC 2025–2026Billion-dollar bets on people and contrarian theses. Zero revenue, zero products.
Built ChatGPT, DALL-E, voice mode. Now building multimodal agentic AI.
One mission: safe superintelligence. No products, no distractions.
LLMs hit a wall. Building world models that learn from reality, not language.
AI as connective tissue for human collaboration.
Novel RL for superintelligence. Three months old.
Open frontier lab. Western answer to DeepSeek. No model shipped yet.
Opening the black box. AI interpretability.
Spatial intelligence. 3D world models from images. ~30 people.
Caveat: Silicon Valley has been here before. Massive pre-product rounds sometimes build category-defining companies — and sometimes they don’t. Thinking Machines Lab lost its CTO and cofounders back to OpenAI within six months of its $2B seed. H Company (ex-DeepMind, $220M seed) lost 3 of 5 cofounders to “operational differences.” SSI’s Daniel Gross left for Meta. xAI lost all 11 cofounders by March 2026. The talent that makes these bets valuable is also the talent most likely to leave. The bet is real. So is the risk.
Every modality now has its own model race. The frontier isn’t just LLMs anymore.
Benchmarks are dead. Meta admitted Llama 4 was tuned specifically to score well on benchmarks — prompting a credibility crisis across the leaderboard. Models are now optimized for benchmarks rather than tested by them. The industry needs new evaluation methods: real-world task completion, user preference studies, and domain-specific assessments.
Source: Meta Llama 4 controversy (TechCrunch, Apr 2025), Scale AI SEAL benchmark initiativeFrontier intelligence is getting radically cheaper. The cost per million tokens for GPT-4-class performance has fallen from $37.50 in 2023 to $0.14 in 2025 — a 99.6% decline in two years.
Take last year’s frontier model: a model that performs similarly on benchmarks is now 500–700× cheaper to run. This collapse in inference cost is what enables the application layer explosion above — and why agentic workflows (which require 10–100× more tokens) are suddenly economically viable.
Source: Epoch AI (Mar 2025), a16z price index, OpenAI/Anthropic/DeepSeek pricing pages, DVC analysisBase generation is commoditizing. Value migrates to orchestration, inference optimization, and proprietary data.
As raw intelligence gets cheaper, the center of gravity shifts. Value moves from generating answers to getting work done.
AI is moving from a system you consult to a system that acts. That changes software from a tool for humans to a layer of labor that can execute across workflows. The agent ecosystem alone has already created 67,000+ engineering openings globally — more than at any point in three years.
~$420M ARR
Answer engine → agentic platform
Transforms a spare Mac mini into an always-on AI agent that controls apps, browses the web, manages files. 19-model orchestration. Personal Computer product launched Feb 2026 — the first consumer device-as-agent play from a search company.
Source: Perplexity / The Verge, 2026$2.5B+ run-rate
9-month ramp to billion-dollar product
GA May 2025 → $2.5B run-rate by Feb 2026. Weekly active users doubled since Jan 2026. Business subscriptions quadrupled. Terminal-first agentic coding drove Anthropic to $380B valuation and $30B Series G.
Source: Anthropic Series G announcement, 2026Acquired by Meta — $2B+
General-purpose agent → Meta's AI backbone
Founded in China, moved to Singapore. $100M ARR in 8 months, $125M+ run-rate at acquisition. 147T tokens processed, 80M+ virtual computers. Microsoft tested in Windows 11. Meta acquiring to power agents across Facebook, Instagram, Meta AI. China investigating potential export law violations. Continues operating independently from Singapore.
Source: CNBC / WSJ / AP News, Dec 2025$2B+ ARR
Revenue doubled in 3 months
Fastest SaaS growth curve in history. ~60% revenue now from enterprise (was individual-first). 50,000+ enterprises, 100M+ lines of enterprise code per day. $500M ARR Jun 2025 → $1B late 2025 → $2B Feb 2026.
Source: Bloomberg / TechCrunch, 2026$10.2B valuation
67% PR merge rate — autonomous SWE
First fully autonomous software engineer. Devin 2.0 handles async multi-step tasks: reads codebase, plans approach, writes code, runs tests, submits PRs. Moving from code completion paradigm to autonomous project execution.
Source: Cognition, 2026"As big as HTML, as big as Linux"
The OS for personal AI — Jensen, GTC 2026
Jensen Huang (GTC 2026): "Every company needs an OpenClaw strategy. This is as big as HTML, as big as Linux." Called it the fastest-growing open-source project in history, surpassing Linux's early adoption pace. Open-source personal AI agent running on Mac mini ($599), RTX PCs, DGX Spark, DGX Station, or cloud VPS (~$30/mo).
Source: NVIDIA GTC 2026 keynote / Business Insider / TechRadar, Mar 2026Enterprise AI Agent Layer
NVIDIA's enterprise wrapper for OpenClaw
NemoClaw (NVIDIA, GTC 2026): Enterprise security layer — network guardrails, privacy router, sandboxed execution via OpenShell. Installs with a single command. Adds Nemotron models + Dynamo inference engine. Jensen's pitch: "OpenClaw for everyone, NemoClaw for the enterprise."
Source: NVIDIA GTC 2026 keynote / TechRadar, Mar 2026Jensen Huang declared the arrival of the "inference inflection point" at GTC 2026: two exponentials colliding — demand for inference growing exponentially while cost per token falls exponentially. The question is no longer whether agents can work. It's whether the business models can sustain them.
We unpack the full business model problem — pricing paradigms, margin squeeze, and why the economics are still unresolved — later in the presentation.
Source: NVIDIA GTC 2026 keynote / Axios / CNBC, Mar 2026"Ability to make software will be a human right soon, and it's not going to feel like making software."— The Vibe Coding Thesis
AppDirect: Non-technical marketing team vibe-coded 200K+ lines of code, built 11 projects with 4 in production, and have 80+ applications in progress across Sales, Finance, HR, and Operations.
Zero-code founder: Built a transcription platform that reached 80,000 users, 1M+ minutes processed, and six-figure ARR — in four months.
Source: Lovable / AppDirect case study, Replit / Whisper AI, 2025Agents building agents: 100% of Claude Code is written in Claude Code. 100% of Perplexity Computer is written in Perplexity Computer — as is this presentation.
The nature of code itself is changing. Humans write abstractions — functions, classes, design patterns — primarily so other humans can read and maintain the code. AI does not need that. It can generate and re-generate from scratch faster than it can navigate a complex abstraction hierarchy. Code is becoming a throwaway artifact rather than a maintained asset. 42% of all committed code is now AI-generated (SonarSource 2026). Projected: 65% by 2027.
Source: SonarSource State of Code 2026But the ceiling is rising faster than the floor. While vibe coding democratizes building, advanced practitioners are diverging fast. Anthropic’s 2026 Agentic Coding report: “Software development is shifting from writing code to orchestrating agents that write code.” Engineers now run multiple AI agents in parallel on one codebase (Vibe Kanban, AutoForge), each on isolated git worktrees, with visual kanban boards for task management. The new SWE job: decompose tasks, spin up agents, review their PRs, resolve merge conflicts. A tech lead managing a team of AI juniors. One company deployed 800+ AI agents internally.
Source: Anthropic 2026 Agentic Coding Trends ReportHow fast does open source move? This week, Anthropic accidentally shipped Claude Code’s entire source code in an npm package — 512,000 lines of TypeScript. Within hours, the community had archived it to GitHub (41,500+ forks), rewritten the core in Python, Go, and Rust, and started building on it. The leaked code revealed an “Undercover Mode” that hides the fact AI is contributing to open-source repos — and an unreleased autonomous agent mode called KAIROS.
Meanwhile, OpenClaw — the open-source Claude wrapper with 247K GitHub stars that Jensen Huang compared to Linux — was systematically killed by Anthropic in four weeks: trademark warning, OAuth blocked, features cloned, then “Channels” absorbed the last differentiator. The creator left to join OpenAI.
The moral: proprietary code is a temporary state. The moment it touches the open internet, the community absorbs, rewrites, and surpasses it before you can issue a takedown. This is the same force that made DeepSeek catch up to GPT-4 in months, MCP hit 82K stars, and models commoditize. Open source does not just compete with proprietary anymore — it metabolizes it.
Source: The Register, Hacker News, GitHub, Mar 31, 2026ChatGPT Agent, Devin, Manus
Subscription / usage-based
Internal infra, VPS, hybrid
Higher ops, better governance
Cursor, Claude Code, Copilot
Software subscription only
Perplexity Mac mini, OpenClaw, DGX Spark
$599+ one-time + low variable
FASTEST GROWINGCloud held 81.1% of agent market share in 2025 — but local-first is the next visible deployment wave
Source: Mordor Intelligence, 2025You must make yourself vulnerable to extract value — but that will change.
Snyk ToxicSkills: 37% of OpenClaw community skills contained flawed code. 200+ GitHub security advisories.
EchoLeak attack on browser agents. Slack AI data exfiltration. Gemini memory poisoning demonstrated.
Agents need files, email, calendar, purchases to be useful. Every permission granted is an attack surface.
Today: accept risk to capture value. Tomorrow: agent-specific security layers, capability-based permissions, cryptographic identity.
Agents now see your screen. Claude Computer Use (Mar 2026) lets Claude see your desktop, launch apps, browse the web, and fill spreadsheets. GPT-5.4 has native computer-use. OpenClaw brought it to open source. Agents no longer need a custom API for every tool — they operate any software through the same interface you do. That is a massive unlock for automating legacy systems that will never get an API.
Source: Anthropic, OpenAI, Mar 2026Agent scans kids' school emails. Finds early dismissal — short day today.
Sends iMessage to nanny: "Short day — pickup at 12:30 instead of 3."
Cleaning lady texts via Telegram: "You're out of garbage bags."
Agent orders from Amazon using its own account and crypto card. No human needed.
Checks weather forecast, pre-cools house for afternoon heat via HVAC.
Adjusts lighting, unlocks door via HomeAssistant, confirms to parent.
Reviews tomorrow's calendar, preps grocery list, charges batteries at off-peak rates.
It’s not just tech companies. A roofing company is using AI agents to pull satellite imagery, cross-reference hail damage patterns, and feed warm leads to their sales team. They’re roofers — not engineers, not a startup. When a roofing company runs AI agents, every company runs AI agents.
Source: @RoundtableSpace / Startup Ideas Podcast, Apr 2026CHEAPER AI ≠ LESS SPEND. Cheaper inference = more workflows clear the ROI threshold.
Task completion sounds simple until you look under the hood. What appears to be one product is really a new software stack in disguise.
Every agentic product — from Cursor to Harvey to Glean — is built on the same fundamental layers.
Click any layer to explore tools, companies, and key data. Hover any company for details.
Agent ↔ Tools / Data
"The REST of AI" — how agents access external tools and data.
Anthropic-led. Adopted by OpenAI, Google, Microsoft.
Agent ↔ Agent
How agents collaborate across organizations.
Google-led. Salesforce, SAP, ServiceNow.
Agent ↔ User
How agents surface work to humans. CopilotKit-led open protocol.
Supported by LangGraph, CrewAI, Microsoft, Google, AWS.
"Together, these three protocols are creating an interoperable agent ecosystem — the TCP/IP moment for AI agents."
The enduring advantage in agents will not come from having a model. It will come from orchestrating the full system around the model: memory, tools, workflows, reliability, and distribution. We have moved from prompt engineering to context engineering.
That architecture is the map of defensibility. The winners will not be the loudest at the frontier; they will be the ones who turn intelligence into dependable, repeatable execution.
If today's economics are strained, the next interface may rewrite them. The moment software starts transacting with software, the market changes shape again.
We've built agents that talk to humans. The next frontier: agents that discover, hire, pay, and supervise each other.
Moltbook — a Reddit-like social network for AI agents — went from niche experiment (Jan 2026) to Meta acquisition (March 10, 2026) in ~6 weeks. Agents posted, commented, upvoted, and gossipped about their human owners.
Probably not. The durable market looks less like "bots posting on bot Reddit" and more like a programmable service economy — authenticated agents discovering each other, negotiating work, moving money, and leaving auditable trails.
Agents can't open bank accounts, pass MFA, or handle card fraud flows. But they can hold programmable balances and transact instantly.
The "body" isn't a humanoid robot — it's a dedicated machine running 24/7 with local files, apps, and persistent memory.
The Mac mini is emerging as the default "agent hardware" — cheap, quiet, always-on, with enough local compute to be a persistent digital worker.
The agent-to-agent economy is real enough to invest in, but early enough that the biggest winners may not be the agents themselves — they may be the companies that provide the protocols, identity, payment rails, and trusted execution environments that let agents safely discover, hire, pay, and supervise one another.
That software stack still runs on steel, silicon, and electricity. The more capable AI becomes, the more brutally physical the system underneath it gets.
Every breakthrough at the application layer is paid for in chips, data centers, cooling, and power. AI is driving the largest infrastructure buildout since the interstate highway system, with capital racing ahead of certainty. The bottleneck is no longer just compute; it is whether the physical world can support the pace.
2026 CapEx will consume ~94% of operating cash flows — vs a 10-year average of 40%. For the first time, hyperscalers collectively hold more debt than cash.
Purpose-built AI infrastructure at hyperscale
Source: Nebius, 2026GPU cloud built for AI-first workloads
Source: CoreWeave S-1, 2025Jensen's core reframe: data centers aren't storage facilities anymore. "Electrons go in, tokens come out." The $700B CapEx sprint is building the world's first generation of AI factories — purpose-built for inference at scale.
FY2026 Revenue — +65% YoY
$1T+ in Blackwell + Vera Rubin orders through 2027
Source: NVIDIA GTC 2026 keynote / CNBC / Axios, Mar 2026Groq LPU → acquired by NVIDIA ($20B tech license, Dec 2025). Now powers NVIDIA's inference architecture.
China generates 2× more electricity than the US and added 543 GW of new power capacity in 2024 alone — more than the US has built in its entire history. By 2030, China is projected to have ~400 GW of spare capacity, triple the power demand of the entire global data center fleet. Energy experts who visit China describe power availability as a "solved problem."
Yet the US consumes nearly 2× more data center electricity. The asymmetry: the US has the chips but is hitting energy bottlenecks (Morgan Stanley forecasts a 44 GW shortfall by 2028). China has the electrons but is constrained by US export controls on high-end GPUs.
The race for AI supremacy may not be won by who builds the best model — but by who solves their bottleneck first.
Source: Brookings (Feb 2026), Fortune, IEA, Morgan Stanley$500B commitment
Joint venture with SoftBank, Oracle, OpenAI
Delays and scope adjustments
Power procurement bottlenecks
1.2GW Abilene campus
First phase operational
As AI shifts from training to inference (Jensen's "inference inflection point"), power demand doesn't decrease — it redistributes. Training clusters run in bursts; inference runs 24/7. Always-on inference = always-on power demand.
Source: NVIDIA GTC 2026 / industry analysisDispatchable MW with an executable timeline is the scarce input.
Once autonomous systems can coordinate digitally, the next step is obvious. They move out of chat windows and into the real world.
"The moment when a robot can do everything better than a human doesn't come once in a decade or once in a lifetime, it happens once in the history of humanity, and we're close to it"— Andrew Wooten, CPO, Rhoda AI
Foundation models powering physical AI
The same transformer architecture powering ChatGPT is now learning to control physical robots
Autonomous vehicles are no longer a concept — they're on the road
$16B raised (largest AV round ever). Expanding from 10 → 20+ cities including Tokyo & London in 2026. 127M autonomous miles. Target: 1M rides/week by end of 2026.
Austin launch June 2025. FSD: 1 collision per 5.3M miles vs national avg 1 per 660K. Cybercab production 2026, <$30K by 2027. Fully driverless tests began Dec 2025.
First commercial driverless permits (Aug 2022). Operates in Wuhan, Chongqing, Shenzhen. Expanding to Abu Dhabi.
Purpose-built robotaxi with no steering wheel. Testing in SF, Vegas, Foster City. Las Vegas as first commercial market.
IPO'd. Operates in Shenzhen, Shanghai, Beijing, Guangzhou.
IPO'd. ~150 cars across Abu Dhabi, Dubai, Riyadh. Middle East expansion.
$750M Series C (Jan 2026). Uber partnership for 25,000 robotaxis.
$1.2B Series D (Feb 2026). SoftBank and NVIDIA backed. End-to-end learned driving.
Nebius subsidiary (ex-Yandex SDC). Live robotaxi on Uber in Dallas. Delivery robots on Uber Eats in 3 cities. Building both AV and last-mile delivery.
8× safer than human drivers — 1 collision per 5.3M miles Source: Tesla Safety Report, 2026
Self-driving trucks are commercially hauling freight on US highways. The $1T US trucking industry is the first autonomous market generating real contracted revenue.
1,000-mile Fort Worth→Phoenix route. Partners: Volvo, PACCAR, FedEx, Uber Freight, Werner. Targeting ~$1B rev by 2030.
First US company with fully driverless trucks at commercial scale (Jan 2026). Fortune 50 retail customers. Partners: Isuzu, NVIDIA, Ryder.
Interstate runs from Texas hub. Customers: J.B. Hunt, Werner Enterprises. Also developing autonomous defense vehicles for US military.
Global labor market by sector — and what's automatable
The infrastructure buildout is breathtaking. But even with $700B flowing in, the fundamental economics of AI are still being figured out.
AI is already reshaping how industries operate, compete, and ship product. Yet many of the companies building the core technology are still burning cash to deliver that transformation. The demand is undeniable; the economics are still unresolved.
OpenAI has 900M weekly active users but only 50M subscribers — a 5.6% paid conversion rate. Compare the ad revenue per user:
OpenAI's $25B ARR blends to ~$28/user across 900M WAU — but 94% are free. Ad revenue alone is estimated at $500M–$800M in 2026, or <$1/yr per free user — a fraction of what Meta and Google extract. Subscriptions carry the economics; ads alone cannot fund inference at scale.
Source: OpenAI (Feb 2026), Statista (Meta ARPU 2025), Reuters (ad estimates), Netflix Q4 2025 earningsThe next trillion-dollar company might run on a business model we haven't seen yet. Subscriptions, usage-based pricing, ads, marketplaces, agent-to-agent payments — AI is still auditioning revenue models, and the winners may not look like any software company that came before.
The default: pay per token consumed. Simple, scalable, commoditizing fast.
Charge for resolved tasks, not raw compute. Aligns cost with value.
Consumer AI tries the ad model. But the math doesn't work — yet.
Reality check: OpenAI’s $100M sounds impressive — until you do the math. That is ~$0.12 per user per year. Google makes ~$60. Fewer than 20% of eligible users see ads daily. AI ads may fund free tiers, but they cannot fund inference at scale.
OpenAI killed Sora — six months after launch. The $1B Disney deal (stock warrants for Mickey Mouse licensing) died with it. IP and licensing proved unsolvable. The biggest AI company is still figuring out which products actually work.
Agentic AI as a Service — Jensen's framing at GTC 2026. Agents don't just answer questions — they complete workflows. Pricing shifts from per-token to per-task. Every SaaS vendor becomes an AaaS vendor — or gets disintermediated by an agent.
Pay per seat → Pay per token → Pay per outcome
Pay per workflow completed. The agent IS the product.
$1.5T+ in professional services mapped by automation readiness. Tap any category for AI contenders.
Outsourced × Judgement — AI augments, humans decide
Outsourced × Intelligence — ripe for full automation
Insourced × Judgement — hardest to automate
Insourced × Intelligence — automation coming fast
The agent-to-agent economy is real enough to invest in, but early enough that the biggest winners may not be the agents themselves — they may be the companies that provide the protocols, identity, payment rails, and trusted execution environments that let agents safely discover, hire, pay, and supervise one another.
More than eight companies are now operating at the frontier. The question is no longer who can build — but who owns the full stack.
Everyone says AI is built on three pillars: Data, Compute, and Talent. They are right — but incomplete. Talent remains the scarcest resource: AI roles are exploding, with a third of all openings concentrated in the Bay Area. Yet the conventional model describes the machinery. It does not describe what makes the machinery work.
Without distribution, the best model is a science project. Whoever controls the surface — search, devices, social, enterprise — chooses which models users touch.
Proprietary loops = moats
Inference overtaking training
74% of startups report inference-dominant costs
$10–20M/yr for top researchers
Best model + slow shipping = loss. Culture is the conversion rate of every other input. Speed of execution is the secret ingredient.
Every giant has unmatched resources — and a critical vulnerability.
Startups win in the seams.
Google processes ~15B queries/day. Perplexity hit 200M daily queries by mid-2025 (~1.3% of Google), up from 30M at the start of the year — targeting 1B/week. ChatGPT handles 2.5B prompts/day across 831M monthly users. Combined AI search share is growing 20%+/month. The sharper risk: AI answer engines reduce high-intent query volume and compress ad inventory economics even if overall share holds.
Google's proprietary TPU stack saves an estimated ~$3B/year vs. third-party compute for AI-augmented search. Ironwood (7th gen TPU) is "the first designed specifically for inference at scale" with 10× compute improvement and 2× power efficiency vs. prior high-perf TPU.
Gemini app: 750M MAU. Google Cloud: 48% growth in Q4 2025, run rate above $70B. All 15 products with 500M+ users now use Gemini models. 8M+ paid Gemini Enterprise seats sold in 4 months. Reuters called Google the AI momentum leader in Feb 2026.
Google invested $3B+ in Anthropic. Anthropic trains and runs Claude on Google TPUs — just expanded to multi-GW deal with Google + Broadcom for 3.5 GW of next-gen TPU capacity starting 2027. Google earns strategic exposure and infrastructure revenue from a frontier lab doing $30B run-rate.
$201B revenue in FY2025, $60.5B net income. 3.58B Family Daily Active People. Meta can stuff AI into Facebook, Instagram, WhatsApp, Messenger, ad tools, smart glasses and creator tools across billions of users — even with a weaker model.
Llama 4 received poor reception (benchmark-tuning controversy). DeepSeek seized the open-weight lead. Meta planned its fourth AI restructuring in six months by Aug 2025. Cut ~15,000 jobs (20% of workforce) while simultaneously projecting $115–135B in AI CapEx for 2026 — the clearest signal yet of replacing headcount with compute. Response: $14.3B investment in Scale AI (49% stake, no voting rights). Scale AI founder Alexandr Wang joined Meta to lead the Superintelligence Lab. Separately, Meta recruited Andrew Tulloch (co-founder of Thinking Machines Lab with Mira Murati) — reportedly offering up to $1.5B in compensation over 6 years.
Meta hired Nat Friedman and Daniel Gross to lead the reorg. Brought in talent from DeepMind, OpenAI, Anthropic. Sam Altman said Meta offered OpenAI employees $100M bonuses. The comp war is real but culture stability remains the question.
100M+ MAU across Copilot apps. M365 Copilot drives real ARPU growth across Office, Azure, and GitHub. Multi-model platform (OpenAI, Anthropic, Mistral) gives enterprises flexibility no other cloud offers.
Only 2.4M daily Copilot web visits vs. ChatGPT’s 400M+. Bing AI never broke through. Consumer identity is invisible. The multi-model strategy is a platform strength but a brand weakness.
OpenAI dependency is real: if OpenAI builds its own cloud, Azure loses its biggest AI differentiator. Microsoft hedged by signing Anthropic and Mistral, but OpenAI is ~70% of AI workloads on Azure.
Best on-device inference silicon in the world (M4, A18 Pro). 2B+ active devices. On-device AI processing could be the privacy moat that no cloud company can match.
Apple Intelligence received “mostly underwhelming” reviews. Siri overhaul delayed a full year. Tim Cook “lost confidence” in AI leadership. Switched from OpenAI to Google Gemini for redesigned Siri stack (Jan 2026).
SpaceX acquired xAI in Feb 2026 in the largest merger in history. Combined valuation: $1.25T. SpaceX valued at $1T, xAI at $250B. Planning an IPO around June 2026 targeting ~$50B.
X provides 500M tweets/day as real-time data flywheel. Grok 3: 93.3% AIME 2025. SpaceX generates ~$8B profit (50% margin). Filed plans for orbital AI data centers with up to 1M satellites.
No enterprise AI playbook. Consumer Grok traction unclear vs. ChatGPT/Claude. Orbital data centers are 2–3 years out. But: vertical integration of AI + rockets + satellites + data is unprecedented.
~$1B invested across 50 AI deals in 2024. By 2025: up to $100B committed to OpenAI, $10B to Anthropic, $2B to xAI. Total disclosed ecosystem financing: $33.8B in rounds NVIDIA participated in.
Back demand creators (model labs, apps) while also backing supply-side lock-in (infrastructure, networking, robotics). Foundation models: $20.9B in round sizes. Apps: $5.3B. Cloud/infra: $3.5B. Robotics: $3.3B. Even fusion energy (Commonwealth Fusion: $863M).
Groq LPU acquired for $20B tech license (Dec 2025). Now powers NVIDIA's inference architecture. The message: NVIDIA doesn't just sell the picks and shovels — it's buying the mine operators too.
AWS: $128.7B revenue in 2025 (grew 24% in Q4). Advertising: $68.6B in 2025 (grew 22%). Combined: $197B of AI-exposed revenue. Amazon's ~$200B in 2026 CapEx is described as covering AI infrastructure and robotics as "seminal opportunities."
If AI agents shop for users, they optimize for price/fit/speed — not for Amazon's sponsored placements. Amazon's $17.3B/quarter ad business is directly threatened. Amazon sent legal threats to Perplexity over agentic shopping, and is updating site code to deter outside AI agents.
Rufus AI shopping assistant: 250M+ users, 60%+ higher conversion. Buy for Me: agentic purchasing across 500K products on other brands' sites. 1M+ robots across 300+ facilities. Strategy: internalize agentic commerce inside Amazon-controlled rails.
The stakes rise fast when intelligence gains a body. At that point, this is no longer just a product cycle; it is national strategy.
AI is no longer just a market contest. It is becoming a contest between national systems, supply chains, and spheres of influence. Two rival stacks are taking shape as export controls tighten, sovereign capital accelerates, and strategic autonomy becomes a requirement.
Released Jan 20, 2025 under MIT license. Matched OpenAI o1 on AIME 2024 (79.8% vs 79.2%), MATH-500 (97.3% vs 96.4%). 20–50× cheaper to use.
Broadcom -17.4%, Oracle -13.8%, Marvell -19.1%. Philadelphia semiconductor index -9.2%.
The lesson: DeepSeek did not prove compute no longer matters. It proved that efficiency gains, better architectures, and RL-heavy post-training can narrow the frontier with much less capital than the market had assumed.
R1 scored 79.8% on AIME 2024 (OpenAI o1: 79.2%), 97.3% on MATH-500 (o1: 96.4%), 71.5% on GPQA Diamond (o1: 75.7%), and 2029 Codeforces rating (o1: 2061). Genuinely frontier-adjacent on all key reasoning benchmarks.
Alibaba released Qwen 2.5-Max on Jan 29, 2025, claiming it beat GPT-4o, DeepSeek-V3, and Llama 3.1-405B across the board. Baidu announced Ernie 4.5 would go open-source from Jun 30, 2025 — a direct strategic reversal linked to DeepSeek pressure.
By Sept 2025, the Qwen family comprised 300+ generative AI models, 600M+ downloads, and 170,000+ derivative models globally. The "China open model" wave became real at ecosystem scale.
On January 13, 2025, the US unveiled an AI diffusion rule that split the world into three tiers based on access to advanced AI chips.
Fixed allocation of 49,901 H100-equivalent GPUs through 2027, with a smaller no-license window of 1,699 H100-equivalents.
Initial chip export controls to China
Controls tightened — NVIDIA H100 restricted
DeepSeek proves constraints accelerate innovation
Three-tier diffusion framework — world split into allies, capped, blocked
Bifurcated AI ecosystem solidifying — two dominant stacks plus contested middle
Risk tiers established: unacceptable (banned), high risk (strict obligations), transparency/limited risk (disclosure), minimal (no rules).
Banned AI practices, AI system definition, and AI literacy obligations start applying.
General-purpose AI obligations become applicable. Member States designate national authorities.
High-risk AI (Annex III) and transparency rules (Article 50) start applying. Formal enforcement begins.
High-risk AI embedded in regulated products gets extended transition. All categories fully enforceable.
Penalty scale: Prohibited-practice violations can reach €35M or 7% of global annual turnover — large enough to matter for any company deploying AI in Europe.
ByteDance's Doubao exceeded 100M DAU by Feb 2026, processing 1.9B AI queries during CCTV's Spring Festival Gala. Doubao-1.5-pro (Jan 2025) was priced at only 3–4% of GPT-4.
Alibaba's Qwen 3.5 (Feb 2026): claimed 60% cheaper to run, 8× larger workloads. Baidu's MuseSteamer (Jul 2025) for enterprise video generation. Tencent's Hunyuan + Yuanbao for text, image, and video.
China's Interim Measures for Generative AI Services took effect Aug 2023. By Mar 2025: 346 services filed. By end-2025: 748 generative AI services completed filing and 435 applications registered. Providers must display model name and filing number.
By 2024: 2,027,000 industrial robots in operation, 295K annual installations, 54% of global demand. Chinese manufacturers captured 57% of the home market for the first time. This manufacturing base gives China dense industrial data and a path to embodied AI scale.
The world looks less like a clean Cold War binary and more like two dominant stacks plus a contested middle:
Privileged access to frontier compute. NVIDIA + hyperscalers. Closed-weight leaders at the top, open-weight ecosystem underneath.
Lower-cost/open models, industrial AI, mass consumer distribution. H800-constrained but architecturally innovative.
Tier 2 states bargain for room to maneuver. Export controls force countries to choose alignment.
The UAE example: G42 divested China investments and removed Chinese hardware to stay inside US rules after Microsoft's $1.5B investment (Apr 2024).
Source: CSIS, Reuters, New Lines Institute>90% of advanced AI chips manufactured on a single island
Source: SIA, 2025The race for AI supremacy may not be won by who builds the best model — but by who solves their bottleneck first. The US has the chips but is hitting energy bottlenecks. China has the electrons but is constrained on silicon. Everyone else is choosing sides.
Geopolitics is accelerating capital deployment, not cooling it. That creates real opportunity, and very real mispricing.
These aren’t Silicon Valley startups. This is a roofer in Texas.
Click “The Data” to see what actually happened →
“If AI were killing tech jobs, recruiters wouldn’t be the hottest hire in the market.”
Source: Lenny Rachitsky / TrueUp, 9,000+ companies, March 2026| Company | Valuation | Revenue | Multiple |
|---|---|---|---|
| OpenAI | $730B | $25B ARR | ~29x |
| Anthropic | $380B+ | $30B run-rate | ~57x |
| Perplexity DVC | $18–20B | ~$300M | ~60x |
| Cursor | $29.3B | $2B+ | ~15x |
| xAI pre-merger | $250B | ~$500M ARR | ~500x |
Pricing power depends on model scarcity — which is eroding fast
Distribution advantages can evaporate when models improve
Robotics, bio, energy — capital-intensive, long payoff horizons
| Dot-Com (2000) | AI (2026) | |
|---|---|---|
| Revenue | Minimal / speculative | $25B+ ARR at frontier |
| Enterprise Adoption | Early experiments | 53% of enterprises deploying |
| Infra Spend | Telco capex bubble | $700B+ hyperscaler capex |
| Moats | Weak — eyeballs only | Data, compute, distribution |
The answer isn't irrational exuberance — it's structural. 1.3% of asset managers control 66% of all capital, and they can't write $5M seed checks. So $147T in global AUM funnels into the few AI companies big enough to absorb it.
Top 10 manage $62T — 42% of global AUM. They need to deploy billions per quarter. Early-stage VC is invisible to them.
Source: BlackRock Q4 2025 earnings, Vanguard, UBS, Fidelity filingsAI captured 61% of all global VC in 2025 ($259B of $427B). But within AI, concentration is extreme:
Tap any bar for details. These 8 rounds alone total ~$250B+ — more than total global VC outside AI.
Source: CNBC, Crunchbase, OECD, company announcements 2025–2026The vast majority of investors can't write small, high-risk VC checks. They need to deploy at scale. So they pile into the ~20 private AI companies large enough to absorb $100M+ checks. Insane competition for a tiny number of deployable deals drives late-stage multiples up — which are then inherited by early-stage VCs.
Sovereign Wealth Funds deployed $46B into AI ventures in 2025. Saudi PIF, Abu Dhabi's Mubadala, Singapore's GIC and Temasek are now among the largest AI investors.
Source: EY Global GenAI VC Report, Dec 202597% of AI deal value went to North America. San Francisco Bay Area alone captured $122B — 75%+ of all US AI funding.
Source: OECD, Crunchbase 2025 dataBe skeptical on valuation. Not on AI demand. The top quartile is pulling away. The bottom quartile is standing still. The gap widens every quarter.
Noise in pricing does not change the direction of the market. It makes disciplined positioning across the stack even more valuable.
The entire world's business is being rebuilt. Every industry — search, commerce, legal, healthcare, finance, property, video, code, robotics — is getting a new AI-native entrant. The incumbents have resources, but they can't do everything at once. They have organizational drag, legacy architectures, and cannibalization risk that slows them down.
The opportunity for startups isn't to avoid competing with Google and OpenAI. It's to move faster in the seams they can't fill. Perplexity competes with Google Search. Cursor competes with GitHub Copilot. Higgsfield competes with Sora. In every case, the startup has the advantage of focus, speed, and a willingness to bet the company on a single wedge.
DVC's portfolio is positioned across the stack because the winners won't cluster in one layer — they'll emerge wherever a startup can claim turf faster than an incumbent can defend it.
A portfolio is a point of view made concrete. From here, the only question that matters is where the forces already in motion take us.
These are not predictions pulled from thin air. They are the logical consequences of every trend in this presentation — extrapolated five years forward. The pattern is clear: AI software moves fast, AI hardware moves slower, and AI adoption in large legacy industries moves slowest of all.
"The 2026–2030 AI story is less about whether AI works, and more about whether institutions — constrained by power, regulation, and labor — can absorb ultra-cheap intelligence quickly enough. The winners will have culture that retains talent, distribution that reaches users, and the discipline to invest across the full stack rather than betting on a single layer."
The trends are clear. The question is who acts on them first.
You are not late to a trend.
You are early to a restructuring of the global economy.
The reset is not to learn more AI jargon. It is to see the stack clearly, understand where value is shifting, and act before today's temporary leaders harden into tomorrow's incumbents.
That is where DVC operates: across the system, across the cycle, and with a bias toward the layers that compound as AI moves from breakthrough to infrastructure.
March 2026