Welcome to the Halftime Show.

For our 15th edition, we’re doing something special: a State of the Union on AI as of April 2026.

We’ll skip the technical benchmark deep dives that you can find elsewhere. Instead, we’re going where most don’t: Macro, Infrastructural, and Commercial.

This is a long one and deviates from our regular Thursday format.

Grab a drink, sit back, and enjoy- it’s time to lock the fk in.

PLAY OF THE WEEK
We need to start viewing AI like the NBA

On June 13, 2019, the Toronto Raptors beat Golden State in Game 6 to win their first NBA championship. The greatest upset in NBA history (according to every Canadian). I was there. I'll remember that night for the rest of my life.

How did Raptors, the underdog from the North of the border, made up of Lowry, undervalued his entire career; Kawhi, a one year import from South Cal; Pascal, a kid from Cameroon who didn't start playing ball 'til 16; Ibaka, traded twice and written off; Van Leet, undrafted and unsigned out of college; and Gasol, dubbed the 34 yr old "basketball dinosaur" END the GOLDEN STATE WARRIOR dynasty?

Surely not with the most talent or the biggest budget.

Masai Ujiri, the president, understood the dogma. You don't need five generational talents. You need the right pieces- trades, player development, coaching hires, switchable defensive -  assembled into a system where every part makes the others better.

You wouldn't answer "which is the best team in the NBA" with "Luka Dončić." Yet that's exactly how we evaluate AI- by asking who has the best model… a profound category error.

On December 3rd in Washington DC, Jensen Huang shared his "five layer cake" analogy for AI: Energy, Chips, Infrastructure, Models, and Applications.

We'll use that framework to evaluate where the most significant technological shift in human history is heading - in true Delta style, as if you're the GM scouting a championship franchise.

What we’re gaming out:

  • Energy: the nutrition & conditioning program

  • Chips: training grounds & equipment

  • Infrastructure: coaching staff & team ops

  • Models: players

  • Applications: championships

  • Turning championships into dynasties: how to win the next decade

By the end, you'll understand "AI" better than most in the room.

AKA Nutrition & Conditioning
Layer 1: Energy

In any elite sports program, nutrition and conditioning is the unsexy, invisible work that no one watches- but it's what allows you to go harder, faster, longer.

Energy is the conditioning program of the AI economy. Every calculation performed by a model, every chip, every data demands electricity. Without it in abundance and at economic cost, the rest of the stack can't scale.

It is also the primary bottleneck. Energy is a finite organic resources that takes time to harvest, generate, and scale. In systems terms, it's the "limiting factor" - the slowest component to scale determines the ceiling for everything above it.

Currently, the global energy landscape reveals a structural imbalance in capacity and cost, driven by aggressive government subsidies:

Metric

United States

China

Total Electrical Generation Capacity

~2 TW

3.89 TW (2x US Capacity)

Industrial Subsidy / Cost Differentials

Market Rate / High Permitting Costs

50% Subsidies for Chip Manufacturers; Subsidized Employee Transportation

All in Fab Operating Cost

Standard benchmark

4-8x cheaper

Even more granular:

Deployment Location

Electricity Rate

Annual Power Cost (10,000-GPU cluster)

China (subsidized industrial)

$0.04/kWh

Baseline

U.S. (standard market)

$0.16/kWh

+$8M+ annually

{ energy abundance at low cost = feasibility for AI scale }

At the Nvidia GTC last month, Jensen Huang rebranded data centers as "AI factories" (more on this later). The product is tokens, and the primary input cost is energy.

Energy cost differential creates a strategic tradeoff: inferior chips become viable through volume deployment, changing the game from 'best chip' to 'most system throughput per dollar.

{ energy arbitrage: $0.04/kWh vs $0.16/kWh = 4x chip deployment at same OpEx }

A chip that draws more power than Nvidia's is a liability in high cost energy markets-but becomes viable at volume when energy is subsidized.

Deploying 4x more chips compensates for per unit efficiency gaps and enables faster infrastructure buildout: more fabs, more data centers, more models, more applications.

Real example: Huawei's Ascend 910C delivers 80% of H100 computing power, consumes significantly more energy per operation, remains economically viable due to subsidized power.

More power is coming

In 2025, China invested $500B in energy projects, an 11% increase YoY.

This is industrial capacity expansion designed to remove energy as a constraint on AI.

(Credit: Reuters)

For context: China added more renewable energy capacity in 2025 than the entire U.S. grid added from all sources. The 15th Five Year Plan (2026–2030) projects 200–300 GW of new renewable capacity annually-sustaining this pace for the next FIVE years.

AKA Training Grounds & Equipment
Layer 2: Chips

Energy keeps the lights on. Chips are the training ground, the weight rack, the tools that turn raw capacity into performance. The game is system level throughput: how much compute you can deploy at what total cost.

But if the best training facility in the world bans you from entry, what do you do?

You build your own gym. Not an Equinox or Gold's at first, but you run it 24/7, and figure out how to make ten athletes training together perform like one elite unit.

Nvidia, as we all know, obliterates the semiconductor layer in per chip efficiency. Nvidia also owns the software ecosystem built around them. CUDA has twenty years and hundreds of millions of installations. Switching costs compound with every model trained.

The H100 and Blackwell architectures set the performance standard globally, and export controls were designed to keep it that way.

(Credit: Getty Images)

{ the assumption: cutting edge lithography restrictions + high end chip export bans = China's AI ambitions stall }

The assumption was largely wrong.

Export controls didn't halt development- they incubated Nvidia's challenger through vertical integration at speed (we wrote about how).

Scale over edge

When you can't access EUV lithography for 3nm chips, you architect around the constraint. And the current answer?

The SuperPod: thousands of lower performance chips interconnected to function as a single logical machine, making up for what each other lacks.

In technical terms, Huawei's Atlas 950 binds 8,192 Ascend chips with 16 PB/s interconnect bandwidth (10x global internet peak), creating a unified system that learns, thinks, and reasons as one. Like a mini data centre.

(Credit: SCMP)

{ 80% performance per chip × 5x volume × 25% operating cost = competitive parity at lower total system cost }

A type of architectural innovation that turns volume into a moat.

Production trajectories demonstrate a rapid closing of the gap:

  • Huawei (HiSilicon): plans to produce 1.6 million Ascend AI accelerators by 2026

  • SMIC: Foundry capacity expanded 15% in 2024, 14% in 2025, with strategic focus on ramping 28nm lines across multiple cities

  • Mature Node Expansion: China based chipmakers projected to account for nearly half of all new global mature node capacity over the next 3-5 years

The capital is flowing to volume production, not frontier research. 19 semiconductor firms initiated Hong Kong IPOs in 2025, with 30 more in A share pipeline targeting $13.6B combined fundraising.

There are two games being played:

  1. Nvidia: per chip performance leadership and ecosystem lock in through CUDA

  2. Challenger/alternative: volume at scale, enabled by cheap power, architectural workarounds (like SuperPod) and fast deployment

We watch both with equal excitement.

All said and done, silicon capability remains theoretical without the physical velocity to deploy it at scale. That's where infrastructure enters.

AKA Coaches & Team Ops
Layer 3: Infrastructure

You can have the best QB and the perfect playbook. But if your coaching staff can't install the system faster than the opponent adjusts their defense, you lose.

Infrastructure is how fast you go from blueprint to deployment. Data centers. Power grids. Manufacturing lines. Supply chains. The physical foundation that determines whether your AI runs in a lab or at scale across millions of users.

Speed of installation beats brilliance of design, especially when AI democratizes output quality.

China's deployment velocity is the no huddle offense of global tech. This phenomenon is what we're most interested in at LOD.

"If you want to build a data centre in the states, probably about 3 years. They can build a hospital in a weekend." - Jensen Huang on Chinese speed.

Without going too deep, three decades of supply chain concentration built the Pearl River Delta into the densest hardware ecosystem on earth. Every component supplier, assembler, and testing facility operates within a 100 kilometer radius. Vertically integrated campuses and streamlined permitting accelerate execution further.

The result is the infamous "72 hr prototype to production" cycle.

{ 72 hour prototyping = zero warehouse rent, zero shipping lag }

Non exhaustive, but indicative illustration below-

The humanoid robotics sector illustrates this tempo at its best:

  • Guangdong's new humanoid robot production line outputs 10,000 units annually, one robot completing assembly/30 mins

  • Agibot scaled from 5,000 - 10,000 units in three months: the fastest production ramp in the sector globally

  • UBTech partnered with Siemens to reach 10,000 unit annual capacity in 2026

Infrastructure velocity = your robots can run half marathons for fun

The infrastructure layer determines whether silicon capability remains theoretical or becomes deployed at billion user scale.

This is something you truly have to see with your own eyes to believe. It's what we witness everyday, here in Shenzhen.

AKA The Players
Layer 4: Models

(Credit: Sports Illustrated)

Now that you have the conditioning, equipment, and coaching dialled in, the talent is ready. AI models are the players. The talent you develop ontop.

They get trained, refined, benchmarked, scouted by enterprises, deployed into production.

Different strengths, different specializations, different cost profiles. Some are franchise stars with massive contracts. Others are high value free agents available to anyone willing to sign them. And some, are unassuming rookies who will become generational talent.

All layers matter, but models are where most leverage is built.

The US leads in frontier model development, bar none. OpenAI's GPT-5.2, Anthropic's Claude Opus 4.6, Google's Gemini 3- these set the performance ceiling. They're also closed and controlled.

When U.S. export controls landed in October 2022, cutting off access to Nvidia's H100 and A100 chips, China faced a choice: wait for domestic chip alternatives to catch up, or architect around the constraint entirely. They chose the latter.

Open source became a national ideology- not out of charity, but necessity.

{ open weights + priced at cost + global developer contributions = ecosystem capture at speed closed APIs can't match }

The AI+ Initiative (August 2025) made it official: government subsidies for open source development, explicit directive to "promote open source AI ecosystems." Today, more than 90% of Chinese enterprises use open source technologies.

"I'd say there's an 80% chance they're using a Chinese open source model," notes Martin Casado, partner at a16z, on startups applying for VC funding.

Two Distribution Strategies

Closed source approach:

  • Maintain proprietary control over model weights

  • Price API access at premium ($3-15 per million tokens)

  • Monetize immediately through inference revenue

  • Protect competitive advantage through secrecy

Open weight approach:

  • Release weights publicly under permissive licenses (MIT, Apache 2.0)

  • Price inference near marginal cost ($0.14-0.46 per million tokens)

  • Accept short term revenue sacrifice for distribution velocity

  • Build switching costs through ecosystem depth

What happened next…

(Source: USCC)

Model

Type

Cost per Million Tokens

Performance Index

Kimi K2.5

Open weight

$1.20

47

GPT-5.2

Closed source

$4.81

47

{ same task (50,000 documents/day) = $4,200/month (US closed models) vs. $210/month (Chinese open weight) = 20x cost gap at performance parity}

Side note, this isn't just enterprise. Take a look at this video created by ByteDance's Seedance 2.0 video gen tool, that caused cease & desist letters from Disney, Paramount, Sony, Warner Bros., Netflix, and MPA within 24 hours of release.

Incredible.

Model

Provider

Monthly Subscription

API Pricing (per second)

Seedance 2.0

ByteDance

$9.60 (~69 RMB)

$0.017-$0.13/sec

Sora 2

OpenAI

$200 (ChatGPT Pro)

$0.10/sec

Veo 3.1

Google DeepMind

~$250 (Google AI Ultra)

$0.10–$0.75/sec

When DeepSeek R1 was trained for ~$6 million, matching GPT-4-class performance at 1/20th the cost, it proved frontier level intelligence could be commoditized.

Within six months, a wave of models non-Chinese developers had never heard of debuted to market. Moonshot's Kimi K2.5. Alibaba's Qwen. MiniMax. Zhipu's GLM. All open weight & priced near cost. By November 2025, Chinese models held 7 of the top 10 download spots on Hugging Face.

(Source: Interconnects AI)

Across the broader landscape:

Model Origin

Type

Average Token Cost (per million)

United States

Closed Source

$3.00 – $15.00

China

Open Weight

$0.14 – $0.46

Distribution. Distribution. Distribution

The adoption numbers tells us whether open source as a distribution tactic worked:

  • 1.2% → 30%: Chinese open weight model share of global usage, late 2024 to end of 2025

  • 80%: estimated share of US startups raising, built on open weight Chinese models (a16z)

  • 100,000+ derivatives: Qwen family ecosystem on Hugging Face- largest model ecosystem globally, exceeding Meta's Llama

  • 700M downloads: Qwen model family total downloads

  • 7 of top 10: Chinese models among most downloaded on Hugging Face (Nov - Dec 2025)

Each derivative represents engineering hours, training data, and product architecture invested into the ecosystem. Switching costs accumulate with every deployment.

And finally, the validation from the big boys:

  • Airbnb relies heavily on Qwen: CEO Brian Chesky called it "very good, fast and cheap"

  • Chamath Palihapitiya: migrated Social Capital workloads to Moonshot's Kimi K2

  • Cursor and Cognition (Devin AI): both speculated built on Chinese base models

  • AWS, Azure, Google Cloud: all serve DeepSeek and Qwen through their enterprise platforms

The cost advantage unlocked application depth that wasn't economically viable before.

When you can run the same intelligence for 3-5% of the price, integration becomes pervasive rather than selective.

By late 2025, the model layer is starting to flatten. Competitive advantage migrated to deployment and execution.

AKA Championships
Layer 5: Applications

(Credit: Toronto Star)

Back to the Raptors. Championships are what you play for. Everything else- the training, the trades, the system- exists to win rings.

That's the Applications layer. Everything built in the layers below…energy, chips, infrastructure, models….only matters if it translates into applications people actually use. Applications are the only layer that generates revenue.

It’s also the only layer that creates the data flywheel feeding everything below it:

{ better applications → more users → more data → trains better models }

In the rest of the world, AI is largely a destination -you go to ChatGPT, you open a separate tool.

In China, AI is plumbing. It runs inside ecosystems that already command hundreds of millions of users, in what's known as "super apps".

The "Super App" Phenomenon

Super apps exist in China due to a combination of technological, economic, social, and regulatory vectors… but most notably, China leapfrogged the PC era completely, and came online via smartphones, creating demand for a single platform that handled everything rather than juggling multiple specialized apps.

Vs the rest of the world who were already locked into specialized, OS level, fragmented apps.

For example:

{WeChat = iMessage + Venmo + Uber + DoorDash + Slack + Instagram, all running in one app with 1 billion+ users and AI embedded across chat, payments, meetings, and documents}

The Aggregate: 2.2 billion monthly active users across the top 100 AI companies in China. Six of the ten largest AI companies by user scale are Chinese:

Company

Analogy

MAUs

Product Portfolio

Baidu

Google + Waymo

730M

22 web products, 5 apps (all AI augmented)

ByteDance

Meta + TikTok + CapCut + GPT-4-level models

372M

21 web products, 13 apps

Meitu

Adobe + Facetune + VSCO + Canva

195M

17 products on shared AI engine

Tencent/WeChat

iMessage + Venmo + Uber + DoorDash + Slack

1B+

AI embedded across chat, payments, meetings, docs (Hunyuan models)

{more users = more data = trains better models = better UX = more users}

Simply put, super apps capture the entire lifecycle of human intent - generating training data across the full user journey that silo'd apps can't replicate.

Going Global

Of the top 23 Chinese AI companies by revenue, 19 generate the majority of their income overseas.

The playbook = "chuhai" (going to sea) = dominate domestically for scale and iteration speed, then export to dollar denominated markets for margin.

What’s next?
From Championships to Dynasties

Winning a championship is hard. Building a dynasty is harder. The difference is whether you've built a system that compounds.

While the first half of the presentation solidified NVIDIA’s dominance in data centers and agentic software, a massive 45 minute segment was dedicated to what Huang calls the "next trillion-dollar wave": Physical AI.

Actually, he put Olaf to work as a demo.

We knew this
The Future of AI is Physical

The era of the chatbot is maturing; the era of Embodied Intelligence is just beginning.

Jensen is betting the roadmap on AI that moves- robotics, autonomous vehicles, and industrial scale deployment:

  • Intelligence: Cosmos aka the "Physical OS" and World Foundation Model that teaches AI the laws of physics

  • Infrastructure: Omniverse aka the digital foundation and physics engine used to build high-fidelity virtual worlds.

  • Application: Isaac Sim aka the specialized robotics sim training platform

  • The Brain: Jetson Thor aka he silicon muscle for humanoids

  • The Scale: RoboTaxi Ready aka a massive ecosystem involving partners like BYD and Hyundai

The 3 Pillars of Physical AI

Deploying intelligence in the real world requires a different toolkit than building a LLM:

  1. Manufacturing Volume: robotics is a hardware game. Iteration velocity is dictated by supply chain depth and production capacity

  2. Deployment Velocity: success depends on "friendly" regulatory environments and subsidized energy to make massive fleets economically viable

  3. The 10% Edge Case: synthetic data gets you to 90%. The final 10%-the life or death edge cases-only comes from actual operational data at scale. (this is what every Chinese firm is chasing)

This is where the two loops described earlier converge with accelerating force:

  • Loop One (Digital): open models → global adoption → derivative ecosystem → rapid iteration → improved base models → wider deployment

  • Loop Two (Physical): deployment → real world data → model refinement → better hardware → more deployment

*Restricting access to cutting edge training chips constrains Loop One (frontier model development). It does not constrain Loop Two (physical data accumulation)

A factory inspection model doesn't need a cluster of H100s; it needs millions of labeled production images. The entity that deploys 10,000 robots owns that dataset.

The Bottom Line: Physical AI is the convergence of digital intelligence and industrial muscle. Deployment is the new Moat.

STAT OF THE WEEK

1

AI is not a single technology.

It's a layered industrial system where each level supports the one above it. A bottleneck at the bottom restricts everything at the top. Interdependencies are absolute.

The next phase of competition extends beyond chat interfaces into factories, roads, warehouses, and production lines- environments where deployment speed, operational data generation, and manufacturing capacity determine who create dynasties.

That’s it for now, team. If you’re still with us, you are now a better GM than 30 minutes ago- and certainly, a greater AI fanatic.

See you next week,
Jen, live from Shenzhen

Keep Reading