David Luan: Why Nvidia Will Enter the Model Space & Models Will Enter the Chip Space | E1169

David Luan: Why Nvidia Will Enter the Model Space & Models Will Enter the Chip Space | E1169

The Twenty Minute VCJun 24, 202458m

David Luan (guest), Harry Stebbings (host), Narrator

Historical phases of AI progress: from bottom-up research to mission-driven scalingTransformers, scaling laws, data vs compute, and new paths to model improvementReasoning, memory, and the emerging split between chatbots and agentsEconomic structure of the model layer: concentration, commoditization, and business modelsVertical integration between chips and models (Nvidia, TPUs, Apple, cloud providers)Enterprise adoption dynamics, RPA vs agents, and workflow automationRegulation, open vs closed models, and human–computer interaction as a missing piece

In this episode of The Twenty Minute VC, featuring David Luan and Harry Stebbings, David Luan: Why Nvidia Will Enter the Model Space & Models Will Enter the Chip Space | E1169 explores aI’s Next Era: Vertical Integration, Smarter Agents, and Chip Wars Ahead David Luan, CEO of Adept and former leader at Google Brain and OpenAI, outlines how AI has shifted from bottom‑up academic research to large, mission‑driven teams solving concrete problems with Transformers as the universal model architecture.

AI’s Next Era: Vertical Integration, Smarter Agents, and Chip Wars Ahead

David Luan, CEO of Adept and former leader at Google Brain and OpenAI, outlines how AI has shifted from bottom‑up academic research to large, mission‑driven teams solving concrete problems with Transformers as the universal model architecture.

He argues that model progress will continue despite talk of diminishing returns, driven first by scaling base models and now increasingly by reinforcement-style loops where models act in environments, generate their own data, and improve reasoning.

Luan predicts a tightly concentrated layer of 5–7 frontier model providers, deep vertical integration between chips and models (with Nvidia moving up-stack and clouds moving down-stack), and a clear separation between creative chatbots and reliable work-focused agents.

He sees the biggest long-term value not in raw models or services, but in vertically integrated agent products that can learn arbitrary enterprise workflows, while warning about regulatory capture, overhyped short‑term expectations, and underappreciated human–computer interaction challenges.

Key Takeaways

AI progress has moved from curiosity-driven papers to Apollo-style, goal-driven projects.

Luan contrasts the 2012–2018 Google Brain era of bottom-up research with OpenAI’s shift to large teams focused on specific big goals (e. ...

Get the full analysis with uListen AI

Diminishing returns to compute are overstated; new training paradigms will soak up vast compute.

Traditional scaling shows predictable gains when you double compute, and now a second frontier is emerging: giving models environments (math tools, theorem provers, notebooks) to explore, fail, and self-generate training data via RL-style loops, which both improves reasoning and demands even more compute.

Get the full analysis with uListen AI

Reasoning will be solved at the model-provider layer through environment-based training, not just more data.

Simply scaling unsupervised internet training can’t teach composition of ideas; Luan expects leading LLM providers to enhance reasoning by training models to act in rich problem-solving environments with feedback, which requires changing the models themselves rather than just fine-tuning on proprietary corpora.

Get the full analysis with uListen AI

Models and chips are on a collision course toward vertical integration.

Clouds need in-house chips for margin and scale advantages (e. ...

Get the full analysis with uListen AI

Agents and chatbots are diverging into distinct product categories with different requirements.

Hallucinations are acceptable or even useful in creative chatbots, but intolerable for agents running real workflows (taxes, logistics); Luan predicts reliable, tool-using, goal-driven agents that operate software on your behalf will develop separately from conversational systems designed for information and companionship.

Get the full analysis with uListen AI

True enterprise value will come from vertically integrated agent products, not generic model APIs.

Luan expects only a handful of frontier model providers, and believes that pure model-seller startups face a short window before commoditization; he positions Adept as an end-to-end agent stack that learns arbitrary workflows from users, akin to a new “system of record for workflows,” which captures more durable value.

Get the full analysis with uListen AI

Regulation and human–computer interaction are underappreciated determinants of AI’s trajectory.

He worries less about over-regulation per se and more about regulatory capture favoring incumbents and choking off open source, and he argues that the field is moving “models first, UX later” when it should start from how humans supervise, correct, and collaborate with increasingly capable systems.

Get the full analysis with uListen AI

Notable Quotes

The next phase of AI after Transformer was not going to be about research paper writing. It was going to be about, ‘Let's choose a major unsolved scientific problem and just try to solve it.’

David Luan

The second way of improving model performance is just starting to be tapped now, and that's also going to absorb a boatload of compute.

David Luan

I actually think agents and chatbots are gonna speciate and turn into two different products.

David Luan

Every enterprise workflow is an edge case.

David Luan (relaying a comment from Parag Agrawal)

I view open really as a way for the rest of the field to keep up with the biggest incumbents, and therefore I think it's actually pretty darn important.

David Luan

Questions Answered in This Episode

How will environment-based training for reasoning (e.g., theorem provers, code sandboxes) practically change how models are built and evaluated over the next few years?

David Luan, CEO of Adept and former leader at Google Brain and OpenAI, outlines how AI has shifted from bottom‑up academic research to large, mission‑driven teams solving concrete problems with Transformers as the universal model architecture.

Get the full analysis with uListen AI

What concrete user interfaces or interaction paradigms might replace simple chat windows as the primary way humans work with powerful agents?

He argues that model progress will continue despite talk of diminishing returns, driven first by scaling base models and now increasingly by reinforcement-style loops where models act in environments, generate their own data, and improve reasoning.

Get the full analysis with uListen AI

In a world where major clouds have their own chips and models, what strategic options remain for independent model labs that don’t vertically integrate?

Luan predicts a tightly concentrated layer of 5–7 frontier model providers, deep vertical integration between chips and models (with Nvidia moving up-stack and clouds moving down-stack), and a clear separation between creative chatbots and reliable work-focused agents.

Get the full analysis with uListen AI

How can regulators avoid both over-restricting innovation and enabling regulatory capture that entrenches today’s largest AI companies?

He sees the biggest long-term value not in raw models or services, but in vertically integrated agent products that can learn arbitrary enterprise workflows, while warning about regulatory capture, overhyped short‑term expectations, and underappreciated human–computer interaction challenges.

Get the full analysis with uListen AI

For enterprises still in the “experimental budget” phase, what criteria should they use to decide when to move from pilots with agents to core, scaled deployments?

Get the full analysis with uListen AI

Transcript Preview

David Luan

What OpenAI realized before basically everybody but DeepMind was that the next phase of AI after Transformer was not going to be about research paper writing. It was going to be about, "Let's choose a major unsolved scientific problem and just try to solve it." The second way of improving model performance is just starting to be tapped now, and that's also going to absorb a boatload of compute. So because of that, I actually am not worried about diminishing returns to compute over time. I think every tier one cloud provider existentially needs to win here.

Harry Stebbings

Ready to go? (instrumental music plays) David, I am so excited for this. I've wanted to do this one for a long time. I've heard so many good things. So first, thank you so much for joining me today.

David Luan

Yeah. Thanks, Harry, for having me on it. I've, um, got to watch some of your cool previous episodes, so it's a, it's a real honor to be on here.

Harry Stebbings

And that's very, very kind of you. I really do appreciate that, man. But you, you've been at some incredible companies as a training ground, so to speak, one of which was Google Brain, and I just wanted to start there. When you think about your biggest takeaways from your time with Google Brain, what would you say ones two are, and how do you think that shaped how you think about building Adept today?

David Luan

Google Brain, um, uh, uh, was, and also now is part of DeepMind, is a really magical place. It's, uh, it's the, I think during the, uh, peak days of AI, AI progress on the research side, right, where every day there was a new paper that came out that would just change the world, that, like, 2012 to 2018 or so era of Google Brain was just, like, incredibly dominant. They did an amazing job picking talent, like the people who invented Transformer, the people who invented the diffusion model, people who did, um, all of these new optimization techniques that we all take for granted today. They were all at Brain at the same time. Like, truly the Bell Labs of, of the era. And I think I learned a lot about how to make, um, how to make pure bottom-up, like to see what good pure bottom-up basic research looks like at Google Brain.

Harry Stebbings

What does that mean, what pure bottoms-up good basic research is?

David Luan

Uh, yeah. So, so I have this worldview, um, of, of AI progress as being a part of a couple different phases, right? And I, I like to think about pre-2012 as basically being pre-history. Of course, um, all of the, uh, OGs in the field would probably not like that if I characterized it that way, but before 2012, like, most of the things we tried just didn't really work, right? Like, you, you had things like, um, uh, a sheep being identified as cats and dogs, and, uh, chatbots that, um, barely said anything coherent, et cetera. But I think, like, there was a p- uh, there was a period between 2012 to, like, 2017 or 2018, where deep learning went from something that, um, people didn't believe in to being, like, the dominant paradigm in the field. Right? And so during that 2012, 2018 era, the way people made progress, um, what I mean by bottom-up basic research is you hire the most brilliant scientists. They come to work every day with, like, no, um, near-term objective they're being held accountable to, um, and they just work together and they think about, you know, "Hmm," like, "I wonder what it'd be like if we could solve this, like, open, uh, technical problem in AI." Like, "How do we go, um, how do we create a model that better understands how to generate images?" And they just go work on that out of their own curiosity and drive and maybe some interest in glory and fame through papers. And they do that for, uh, they, they do that for, like, six months or so, and then out pops out this, like, research paper that gets posted to arXiv and goes to a p- a, a journal that, um, that, like, just solves the problem. Like, that's, that's huge. Right? And so, m- the reason I call it bottom-up is because it's just driven by the natural, uh, interactions between all, all these researchers in a, in a setting, and they figure out what they wanna do.

Install uListen to search the full transcript and get AI-powered insights

Get Full Transcript

Get more from every podcast

AI summaries, searchable transcripts, and fact-checking. Free forever.

Add to Chrome