DeepSeek, China, OpenAI, NVIDIA, xAI, TSMC, Stargate

DeepSeek, China, OpenAI, NVIDIA, xAI, TSMC, Stargate

More

Descriptions:

Lex Fridman brings together Dylan Patel, founder of semiconductor research firm SemiAnalysis, and Nathan Lambert, research scientist at the Allen Institute for AI, for one of the most technically substantive conversations about the global AI landscape to emerge from the DeepSeek moment. The discussion centers on DeepSeek V3 and R1 — open-weight models from a Chinese lab that delivered near-frontier performance at a fraction of the compute cost previously assumed necessary — and what their release means for OpenAI, Google, xAI, Meta, Anthropic, NVIDIA, and TSMC.

The conversation goes deep on mechanism: why DeepSeek’s mixture-of-experts architecture and training efficiencies upend assumptions about compute scaling, how KV caching and the autoregressive nature of token generation explain why output tokens cost roughly four times more to serve than input tokens, and why reasoning models with long output contexts create exponentially harder inference challenges. Geopolitical dimensions — US export controls on NVIDIA GPUs, Taiwan’s role in the semiconductor supply chain, and the strategic implications of open-weight Chinese models — are examined with specificity and nuance.

Patel and Lambert are among the most credible analysts covering AI infrastructure and training, and this episode delivers exceptional density of technical and strategic insight. For anyone tracking the post-DeepSeek competitive landscape, model economics, or the hardware constraints shaping frontier AI development, this is essential viewing.


📺 Source: Lex Fridman
🏷️ Format: Deep Dive

3 Items

Companies