No Priors Ep. 54 | With Sarah Guo & Elad Gil

No Priors Ep. 54 | With Sarah Guo & Elad Gil

No PriorsMar 7, 202442m

Elad Gil (host), Sarah Guo (host)

State of frontier models: Gemini 1.5, Mistral Large, Sora, and context windowsRetrieval vs. long-context reasoning and the evolving role of RAGSpecialized models in biology, robotics, and other scientific domainsAgent architectures, reinforcement learning, and targeted application domainsNvidia’s dominance, GPU upgrade cycles, and hyperscaler AI capexEnterprise AI adoption, automation of services (e.g., customer support), and ROISemiconductor supply chain, second sources to Nvidia, and fab/geopolitics dynamics

In this episode of No Priors, featuring Elad Gil and Sarah Guo, No Priors Ep. 54 | With Sarah Guo & Elad Gil explores aI Models, Chips, and Agents: Google, Nvidia, and Enterprise Disruption Sarah Guo and Elad Gil survey the rapid progress in AI models, including Google's Gemini 1.5, OpenAI's Sora, and Mistral's fast-rising large models, and debate the future of retrieval, long context windows, and agentic systems.

AI Models, Chips, and Agents: Google, Nvidia, and Enterprise Disruption

Sarah Guo and Elad Gil survey the rapid progress in AI models, including Google's Gemini 1.5, OpenAI's Sora, and Mistral's fast-rising large models, and debate the future of retrieval, long context windows, and agentic systems.

They argue Google has reawakened as a serious AI contender, while specialized models in biology, robotics, and other scientific domains are poised to proliferate in 2024–2025.

On the infrastructure side, they discuss Nvidia’s continued dominance, the economics of massive GPU capex, and how hyperscalers drive most AI spend, with Meta as a clear example of ROI from heavy AI investment.

They highlight early but real enterprise impact—like Klarna’s AI assistant replacing the work of 700 agents—as a sign that services, software, and labor markets will be reshaped as AI applications finally catch up with the infrastructure wave.

Key Takeaways

Longer context windows expand, not replace, retrieval-based systems.

Models like Gemini 1. ...

Get the full analysis with uListen

Google has reemerged as a serious AI heavyweight.

Gemini 1. ...

Get the full analysis with uListen

Agents will work first in narrow, feedback-rich environments.

Rather than generic ‘do-anything’ agents, progress is coming from systems embedded in constrained domains (games, code, specific web apps) where reinforcement learning, sampling, and validation are possible and data for post-training can be economically collected.

Get the full analysis with uListen

AI infrastructure spend is led by hyperscalers, and Nvidia’s moat is deep.

Most of the capital flowing into GPUs comes from cloud and big tech (Microsoft, Meta, Amazon, Google, etc. ...

Get the full analysis with uListen

Big-tech AI capex can already show outsized enterprise value creation.

Meta’s multibillion-dollar AI infrastructure investments translated into materially better targeting, recommendations, and tools, helping add nearly $200B in market cap in a single earnings reaction, illustrating that large AI bets can pay off at scale.

Get the full analysis with uListen

Enterprise AI automation is starting to hit services jobs in a measurable way.

Klarna’s AI assistant handled two-thirds of customer service chats across 23 markets, matched human satisfaction, cut repeat queries by 25%, reduced resolution time from 11 to 2 minutes, and did the work of ~700 agents, foreshadowing rapid adoption waves as competitors are forced to follow.

Get the full analysis with uListen

The application wave is only beginning relative to infrastructure progress.

2023 was dominated by models and infra; 2024–2025 should see a proliferation of vertical and workflow-specific apps, as a new wave of founders with domain expertise finally mobilizes around obvious gaps in areas like customer service, professional services, and vertical SaaS.

Get the full analysis with uListen

Notable Quotes

The thing that was lacking until recently was the will. And it seems like now, because of the competitive dynamic, the will has been reborn.

Elad Gil (on Google and Gemini)

I’m more of the belief that [large context] just opens up the set of trade-offs you can make between retrieval, more sophisticated retrieval, and model reasoning.

Sarah Guo

We haven’t seen anything yet really in the app wave… many of the apps so far were started by people who were very close to the research community.

Elad Gil

There’s a joke that the foundation model companies are here to replace all the jobs, but they don’t understand what any of the jobs are.

Sarah Guo

The more I learn, the less I know in AI, and it’s the opposite of every other field I’ve ever been in.

Elad Gil

Questions Answered in This Episode

How will the balance between long-context models and retrieval systems evolve in real-world enterprise architectures over the next few years?

Sarah Guo and Elad Gil survey the rapid progress in AI models, including Google's Gemini 1. ...

Get the full analysis with uListen AI

Which specific domains—like biology, robotics, or customer support—are most likely to produce the first breakout ‘AI-native’ applications with massive impact on labor and cost structures?

They argue Google has reawakened as a serious AI contender, while specialized models in biology, robotics, and other scientific domains are poised to proliferate in 2024–2025.

Get the full analysis with uListen AI

What would a credible second-source competitor to Nvidia actually need to get right beyond raw chip design—ecosystem, software, fabs, or something else?

On the infrastructure side, they discuss Nvidia’s continued dominance, the economics of massive GPU capex, and how hyperscalers drive most AI spend, with Meta as a clear example of ROI from heavy AI investment.

Get the full analysis with uListen AI

How should policymakers and company leaders prepare for scenarios where AI agents replicate Klarna-like productivity gains across many service sectors simultaneously?

They highlight early but real enterprise impact—like Klarna’s AI assistant replacing the work of 700 agents—as a sign that services, software, and labor markets will be reshaped as AI applications finally catch up with the infrastructure wave.

Get the full analysis with uListen AI

At what point do general-purpose frontier models become ‘good enough,’ making differentiated value shift primarily to data, vertical focus, and product design rather than raw model capability?

Get the full analysis with uListen AI

Transcript Preview

Elad Gil

(instrumental music) . Today on No Priors, we're having a special episode of Sarah and me just talking. Hello, Sarah. How are you?

Sarah Guo

Hey, Elade. What's going on? I see you a lot.

Elad Gil

Not much. Good to see you.

Sarah Guo

Let's talk about models. What's going on in the model world?

Elad Gil

Yeah. Um, I guess there's a lot of hand models that are emerging, so I was thinking of maybe trying to do that-

Sarah Guo

Mm.

Elad Gil

... eventually.

Sarah Guo

It's almost as good of a business as, as investing.

Elad Gil

I know, right? Um, yeah, so there's been a lot that's happened in the model world, uh, recently. Obviously, Google launched Gemini, which I think had a few interesting characteristics both in terms of, uh, performance, but also the huge context window, right? It was a million token context window. Uh, companies like Magic, I think, in the past have actually put out, like, a 5 million token context window model and things like that, but it's really exciting to see that. And I think for certain application areas like biology, longer context windows actually seem to be quite important. And so for example, if you're doing a protein folding model and you have a short context window, you're often actually not encapsulating much of the protein, right? The average, uh, protein is, I think, something like 300 amino acids long, at least in the human genome, but there are things that are dramatically larger than that and so you just can't capture it in some of the context windows being used for biological models. And so I do think this is gonna be one of those areas that will end up being more important than people think, at least in the short run. Um, but Gemini 1.5 seems to have some really interesting performance characteristics. There's obviously Sora from OpenAI, which was, um, the video model that, uh, you know, is just beautiful to watch. You know, there's other model companies like Pika and others that I think are doing exciting things as well. And then, um, Mistral or Le Miz launched, uh, Le Chat, which is really the name of the product.

Sarah Guo

Le Big Model.

Elad Gil

Le Big Model. Le Big Mac.

Sarah Guo

I believe they call it Mistral Large. Yes. (laughs)

Elad Gil

Yes. Le Large. Mistral Large. (laughs) They launched that, and the thing that's really, really impressed me about Mistral is just the velocity of shipping. It's incredibly impressive. They went from basically starting the company to almost GPT4 level in less than a year, right?

Sarah Guo

Nine months, yep.

Elad Gil

It's amazing. And they have, uh, you know, small performant models. They have Le Big Mac or, you know, the large model. They have chat. They have multiple languages. It's just- it's very impressive execution. So... And then I think the other thing that they just launched or announced was that deal with Microsoft where, you know, they're- they're now being licensed onto Azure, and so I think the main models in Azure now are OpenAI, LLaMA, Mistral, and then some of the Microsoft models. So again, that's striking as well. So, uh, just very impressive progress by that company so far.

Install uListen to search the full transcript and get AI-powered insights

Get Full Transcript

Get more from every podcast

AI summaries, searchable transcripts, and fact-checking. Free forever.

Add to Chrome