
DeepSeek, China, OpenAI, NVIDIA, xAI, TSMC, Stargate, and AI Megaclusters | Lex Fridman Podcast #459
Lex Fridman (host), Nathan Lambert (guest), Dylan Patel (guest), Guest (guest), Guest (guest), Guest (guest)
In this episode of Lex Fridman Podcast, featuring Lex Fridman and Nathan Lambert, DeepSeek, China, OpenAI, NVIDIA, xAI, TSMC, Stargate, and AI Megaclusters | Lex Fridman Podcast #459 explores deepSeek’s Shockwave: Cheap Chinese AI, Chips, and Geopolitics Collide Lex Fridman speaks with semiconductor analyst Dylan Patel and AI researcher Nathan Lambert about the ‘DeepSeek moment’—China’s release of powerful, ultra‑cheap, open‑weight reasoning models V3 and R1—and why it matters technically, economically, and geopolitically.
DeepSeek’s Shockwave: Cheap Chinese AI, Chips, and Geopolitics Collide
Lex Fridman speaks with semiconductor analyst Dylan Patel and AI researcher Nathan Lambert about the ‘DeepSeek moment’—China’s release of powerful, ultra‑cheap, open‑weight reasoning models V3 and R1—and why it matters technically, economically, and geopolitically.
They unpack how DeepSeek achieved frontier‑level performance at a fraction of OpenAI’s apparent cost, through mixture‑of‑experts architectures, low‑level GPU optimizations, and aggressive post‑training and reinforcement learning for reasoning.
The conversation then zooms out to export controls, NVIDIA’s dominance, TSMC’s centrality, U.S.–China–Taiwan tensions, power‑hungry AI megaclusters like OpenAI’s proposed “Stargate,” and what all this implies for AGI timelines and global stability.
Throughout, they revisit open‑source vs. closed models, the future of agents and reasoning, and how rapidly cheaper, more capable AI will reshape software, industry, and possibly geopolitical power balances.
Key Takeaways
DeepSeek proved frontier‑class, open‑weight reasoning can be ultra‑cheap—and that’s a genuine shock to the ecosystem.
DeepSeek V3 (chat) and R1 (reasoning) match or approach GPT‑4‑class capabilities on many benchmarks while reportedly costing only millions to train, thanks to mixture‑of‑experts, custom attention, and brutal low‑level CUDA‑bypassing optimizations. ...
Get the full analysis with uListen AI
Mixture‑of‑experts and custom attention (MLA) are now central levers for cost‑efficient scaling.
Rather than activating all parameters for every token, DeepSeek’s MoE architecture turns on a small subset of ‘experts’ per step, dramatically cutting training and inference FLOPs while keeping a huge total parameter “knowledge” budget. ...
Get the full analysis with uListen AI
Reasoning models fundamentally change inference economics by exploding test‑time compute.
Models like OpenAI o1/o3 and DeepSeek R1 generate long chains of thought and often sample many candidate solutions per query, especially on hard tasks like ARC‑AGI or complex code. ...
Get the full analysis with uListen AI
Export controls constrain China’s *AI usage* and density more than its ability to train frontier models at all.
Patel and Lambert argue that with 10k–50k+ GPUs already in‑country (and access to H20‑class parts, smuggling, and cloud rentals), focused Chinese teams like DeepSeek can still reach the frontier. ...
Get the full analysis with uListen AI
TSMC remains a single point of failure for global semiconductors—and thus for AI.
TSMC manufactures the majority of advanced chips used in servers, phones, cars, and AI accelerators. ...
Get the full analysis with uListen AI
AI megaclusters and power constraints are becoming the new strategic battleground.
OpenAI’s proposed ‘Stargate’ and similar projects at Meta, xAI, Amazon, and Google envision multi‑hundred‑thousand‑GPU clusters drawing multiple gigawatts—on par with or exceeding large cities or nuclear plants. ...
Get the full analysis with uListen AI
Open‑source and open‑weight models will keep advancing, but true openness (data + code + weights) is rare and crucial.
DeepSeek R1’s open weights and detailed paper significantly raise the bar, but training data and full code are still opaque. ...
Get the full analysis with uListen AI
Notable Quotes
“The ‘DeepSeek moment’ is real. Five years from now, people will still remember it as a pivotal event in tech history.”
— Lex Fridman
“Every company that’s trying to push the frontier of AI has failed runs. You need failed runs to push the envelope on your infrastructure.”
— Nathan Lambert
“If you think AGI is five or ten years away, export controls probably guarantee China wins long term—*unless* AI does something transformative in the short term.”
— Dylan Patel
“Open weights means you can download the model to a computer in your own house that has no internet and you’re totally in control of your data.”
— Nathan Lambert
“Superhuman persuasion will happen before superhuman intelligence.”
— Dylan Patel (citing Sam Altman)
Questions Answered in This Episode
If reasoning models make inference so expensive, what kinds of applications will actually justify paying for deep chain‑of‑thought versus using cheaper ‘good enough’ chat models?
Lex Fridman speaks with semiconductor analyst Dylan Patel and AI researcher Nathan Lambert about the ‘DeepSeek moment’—China’s release of powerful, ultra‑cheap, open‑weight reasoning models V3 and R1—and why it matters technically, economically, and geopolitically.
Get the full analysis with uListen AI
How far can export controls realistically slow China’s *effective* AI power if smuggling, cloud rentals, and domestic chip efforts continue to scale?
They unpack how DeepSeek achieved frontier‑level performance at a fraction of OpenAI’s apparent cost, through mixture‑of‑experts architectures, low‑level GPU optimizations, and aggressive post‑training and reinforcement learning for reasoning.
Get the full analysis with uListen AI
At what point do the risks of extremely persuasive, personalized AI agents—especially when aligned to different national or corporate interests—outweigh the benefits of open models?
The conversation then zooms out to export controls, NVIDIA’s dominance, TSMC’s centrality, U. ...
Get the full analysis with uListen AI
Given TSMC’s central role, is it more realistic to diversify advanced manufacturing away from Taiwan, or to focus primarily on securing Taiwan geopolitically?
Throughout, they revisit open‑source vs. ...
Get the full analysis with uListen AI
As AI makes high‑quality software far cheaper to produce, which sectors or professions beyond programming are likely to be most transformed—or destabilized—first?
Get the full analysis with uListen AI
Transcript Preview
The following is a conversation with Dylan Patel and Nathan Lambert. Dylan runs SemiAnalysis, a well-respected research and analysis company that specializes in semiconductors, GPUs, CPUs, and AI hardware in general. Nathan is a research scientist at the Allen Institute for AI, and is the author of the amazing blog on AI called Interconnects. They are both highly respected, read, and listened to by the experts, researchers, and engineers in the field of AI. And personally, I'm just a fan of the two of them. So, I used the DeepSeek moment that shook the AI world a bit as an opportunity to sit down with them and lay it all out. From DeepSeek, OpenAI, Google, xAI, Meta Anthropic, to NVIDIA and TSMC, and to US, China, Taiwan relations, and everything else that is happening at the cutting edge of AI. This conversation is a deep dive into many critical aspects of the AI industry. While it does get super technical, we tried to make sure that it's still accessible to folks outside of the AI field by defining terms, stating important concepts explicitly, spelling out acronyms, and in general, always moving across the several layers of abstraction and levels of detail. There is a lot of hype in the media about what AI is and isn't. The purpose of this podcast, in part, is to cut through the hype, through the bullshit, and the low resolution analysis, and to discuss in detail how stuff works and what the implications are. Let me also, if I may, comment on the new OpenAI o3 mini reasoning model, the release of which we were anticipating during the conversation, and it did indeed come out right after. Its capabilities and cost are on par with our expectations, as we stated. OpenAI o3 mini is indeed a great model, but it should be stated that, uh, DeepSeek R1 has similar performance on benchmarks, is still cheaper, and it reveals its chain of thought reasoning, which o3 mini does not. It only shows a summary of the reasoning. Plus, R1 is open weight and, uh, o3 mini is not. By the way, I got a chance to play with, uh, o3 mini, and uh, anecdotal vibe check-wise, I felt that o3 mini, specifically o3 mini high, is, uh, better than R1. Still, for me personally, I find that Claude Sonnet 3.5 is the best model for programming, except for tricky cases where I will use o1 pro to brainstorm. Either way, many more better AI models will come, including reasoning models, both from American and Chinese companies. They will continue to shift the cost curve. But the "DeepSeek moment" is indeed real. I think it will still be remembered five years from now as a pivotal event in tech history, due in part to the geopolitical implications, but for other reasons too, as we discuss in detail from many perspectives in this conversation. This is the Lex Fridman Podcast. To support it, please check out our sponsors in the description, and now, dear friends, here's Dylan Patel and Nathan Lambert. A lot of people are curious to understand China's DeepSeek AI models, so let's lay it out. Nathan, can you describe what DeepSeek V3 and DeepSeek R1 are, how they work, how they're trained? Let's, uh, look at the big picture and then we'll zoom in on the details.
Install uListen to search the full transcript and get AI-powered insights
Get Full TranscriptGet more from every podcast
AI summaries, searchable transcripts, and fact-checking. Free forever.
Add to Chrome