Skip to content
Lex Fridman PodcastLex Fridman Podcast

DeepSeek, China, OpenAI, NVIDIA, xAI, TSMC, Stargate, and AI Megaclusters | Lex Fridman Podcast #459

Dylan Patel is the founder of SemiAnalysis, a research & analysis company specializing in semiconductors, GPUs, CPUs, and AI hardware. Nathan Lambert is a research scientist at the Allen Institute for AI (Ai2) and the author of a blog on AI called Interconnects. Thank you for listening ❤ Check out our sponsors: https://lexfridman.com/sponsors/ep459-sb See below for timestamps, transcript, and to give feedback, submit questions, contact Lex, etc. *Transcript:* https://lexfridman.com/deepseek-dylan-patel-nathan-lambert-transcript *CONTACT LEX:* *Feedback* - give feedback to Lex: https://lexfridman.com/survey *AMA* - submit questions, videos or call-in: https://lexfridman.com/ama *Hiring* - join our team: https://lexfridman.com/hiring *Other* - other ways to get in touch: https://lexfridman.com/contact *EPISODE LINKS:* Dylan's X: https://x.com/dylan522p SemiAnalysis: https://semianalysis.com/ Nathan's X: https://x.com/natolambert Nathan's Blog: https://www.interconnects.ai/ Nathan's Podcast: https://www.interconnects.ai/podcast Nathan's Website: https://www.natolambert.com/ Nathan's YouTube: https://youtube.com/@natolambert Nathan's Book: https://rlhfbook.com/ *SPONSORS:* To support this podcast, check out our sponsors & get discounts: *Invideo AI:* AI video generator. Go to https://lexfridman.com/s/invideoai-ep459-sb *GitHub:* Developer platform and AI code editor. Go to https://lexfridman.com/s/github-ep459-sb *Shopify:* Sell stuff online. Go to https://lexfridman.com/s/shopify-ep459-sb *NetSuite:* Business management software. Go to https://lexfridman.com/s/netsuite-ep459-sb *AG1:* All-in-one daily nutrition drinks. Go to https://lexfridman.com/s/ag1-ep459-sb *OUTLINE:* 0:00 - Introduction 3:33 - DeepSeek-R1 and DeepSeek-V3 25:07 - Low cost of training 51:25 - DeepSeek compute cluster 58:57 - Export controls on GPUs to China 1:09:16 - AGI timeline 1:18:41 - China's manufacturing capacity 1:26:36 - Cold war with China 1:31:05 - TSMC and Taiwan 1:54:44 - Best GPUs for AI 2:09:36 - Why DeepSeek is so cheap 2:22:55 - Espionage 2:31:57 - Censorship 2:44:52 - Andrej Karpathy and magic of RL 2:55:23 - OpenAI o3-mini vs DeepSeek r1 3:14:31 - NVIDIA 3:18:58 - GPU smuggling 3:25:36 - DeepSeek training on OpenAI data 3:36:04 - AI megaclusters 4:11:26 - Who wins the race to AGI? 4:21:39 - AI agents 4:30:21 - Programming and AI 4:37:49 - Open source 4:47:01 - Stargate 4:54:30 - Future of AI *PODCAST LINKS:* - Podcast Website: https://lexfridman.com/podcast - Apple Podcasts: https://apple.co/2lwqZIr - Spotify: https://spoti.fi/2nEwCF8 - RSS: https://lexfridman.com/feed/podcast/ - Podcast Playlist: https://www.youtube.com/playlist?list=PLrAXtmErZgOdP_8GztsuKi9nrraNbKKp4 - Clips Channel: https://www.youtube.com/lexclips *SOCIAL LINKS:* - X: https://x.com/lexfridman - Instagram: https://instagram.com/lexfridman - TikTok: https://tiktok.com/@lexfridman - LinkedIn: https://linkedin.com/in/lexfridman - Facebook: https://facebook.com/lexfridman - Patreon: https://patreon.com/lexfridman - Telegram: https://t.me/lexfridman - Reddit: https://reddit.com/r/lexfridman

Lex FridmanhostNathan LambertguestDylan PatelguestGuestguest
Feb 3, 20255h 6mWatch on YouTube ↗

FREQUENTLY ASKED QUESTIONS

Direct answers grounded in the episode transcript. Tap any timestamp to verify against the source.

  1. What is the difference between DeepSeek V3 and DeepSeek R1?

    DeepSeek V3 is the fast chat-style model, while DeepSeek R1 is the reasoning version that shows its working. Nathan Lambert explains V3 as the experience most people know from ChatGPT: you ask a question, it quickly produces a polished answer, often in a markdown-style format, across many domains. R1 changes the interaction by generating a long reasoning section first. It breaks down the problem, talks through what it needs to do, and then switches into a final answer that summarizes the reasoning. DeepSeek made this visible to users, which helped the model spread beyond the AI community, because people could watch the model work through a problem. Nathan contrasts that with OpenAI's interface, which summarizes the reasoning process into short status updates before showing the answer.

    19:31 in transcript
  2. Why is DeepSeek R1 so cheap to run?

    DeepSeek R1's low price comes from architecture, serving choices, and comparison against high-margin competitors. Dylan Patel points first to multi-head latent attention, or MLA, as a real architectural innovation that reduces memory pressure compared with standard transformer attention. Nathan Lambert adds that MLA can save about 80 to 90% of memory in the attention mechanism, while cautioning that this does not make the whole model 80 to 90% cheaper. Dylan also separates pricing from actual cost. OpenAI's inference gross margins are described as north of 75%, while other providers serving the same open-weight model still cost roughly five to seven times more than DeepSeek. That remaining gap, in Dylan's view, comes from DeepSeek's legitimate efficiency advantages: MLA, mixture of experts design, and low-level libraries that likely carry over from training to inference.

    2:11:08 in transcript
  3. How do GPU export controls affect China's AI race?

    GPU export controls mainly restrict how much AI China can run, not whether Chinese teams can train models at all. Nathan Lambert says there are not many worlds where China cannot train AI models, because the controls mostly kneecap the amount and density of compute available. DeepSeek V3 is his example of a focused team still reaching the frontier with a 2,000 GPU cluster. The bigger pressure is inference and deployment: a huge AI market may need 100,000 GPUs just to serve ChatGPT-like systems. Dylan Patel makes the same distinction in economic terms, saying that simply training a model does effectively nothing unless the compute exists to deploy it into productivity, military capability, or economic growth. The US cannot cut everything off, so the goal becomes keeping a compute gap.

    1:02:02 in transcript
  4. Why is TSMC so important for AI chips?

    TSMC matters because advanced chip manufacturing depends on a tiny set of R&D centers, especially Hsinchu. Dylan Patel says manufacturing can be distributed globally, but the people improving the next semiconductor processes are concentrated in places such as Hsinchu, Hillsboro, and South Korea. That is why he calls Arizona a paperweight if Hsinchu disappeared: within a year or a couple years, the Arizona fab would stop producing too. The dependence is not limited to elite AI accelerators. Dylan says TSMC chips sit behind servers, GPUs, laptops, phones, vehicles, fridges, and even unglamorous power ICs that convert voltage. Earlier, he explains why the industry moved this way: TSMC's foundry model let chip designers outsource manufacturing as advanced fabs became too expensive and difficult for most companies to build alone.

    1:43:29 in transcript
  5. What is Stargate in the AI megacluster race?

    Stargate is an AI infrastructure joint venture whose headline numbers Dylan Patel treats cautiously. He says the announced $500 billion figure is not money already in hand, and even the $100 billion phase-one number is closer to total cost of ownership than direct investment. The first phase is tied to Abilene, Texas, where Dylan describes a 2.2 gigawatt site with about 1.8 gigawatts consumed. Oracle had already been building the first section before Stargate, and OpenAI later got access through the joint venture. Dylan estimates the first section at roughly $5 billion to $6 billion of server spend plus about another billion in data center spend. Filling the whole site with next-generation NVIDIA chips would be closer to $50 billion of server cost, plus power, operations, maintenance, and rental costs.

    4:48:25 in transcript

Answers are AI-generated from the transcript and may contain errors. Tap a question to verify against the source.

Get more out of YouTube videos.

High quality summaries for YouTube videos. Accurate transcripts to search & find moments. Powered by ChatGPT & Claude AI.

Add to Chrome