
No Priors Ep. 27 | With Sarah Guo & Elad Gil
Sarah Guo (host), Elad Gil (host)
In this episode of No Priors, featuring Sarah Guo and Elad Gil, No Priors Ep. 27 | With Sarah Guo & Elad Gil explores gPU Crunch, AI Agents, And Startup Survival In Early AI Era Sarah Guo and Elad Gil discuss the current GPU shortage, its causes in semiconductor supply chains, and the surge in AI-driven demand that outpaces manufacturing capacity. They explore second-order effects such as new GPU-cloud businesses, opportunities for alternative AI chips, and renewed interest in compute-efficient research techniques. The conversation then shifts to AI agents, arguing that focused, vertical use cases will win over vague, general-purpose assistants, and outlining a framework of product, research, and infrastructure-driven approaches. They close by examining private tech and venture markets, predicting significant fallout for 2021-era unicorns, and advising founders to focus on underlying business health rather than clinging to inflated valuations.
GPU Crunch, AI Agents, And Startup Survival In Early AI Era
Sarah Guo and Elad Gil discuss the current GPU shortage, its causes in semiconductor supply chains, and the surge in AI-driven demand that outpaces manufacturing capacity. They explore second-order effects such as new GPU-cloud businesses, opportunities for alternative AI chips, and renewed interest in compute-efficient research techniques. The conversation then shifts to AI agents, arguing that focused, vertical use cases will win over vague, general-purpose assistants, and outlining a framework of product, research, and infrastructure-driven approaches. They close by examining private tech and venture markets, predicting significant fallout for 2021-era unicorns, and advising founders to focus on underlying business health rather than clinging to inflated valuations.
Key Takeaways
Expect persistent GPU bottlenecks as AI demand outpaces physical chip manufacturing.
With NVIDIA far ahead on high-end GPUs, limited foundry capacity, and specialized tooling constraints, supply cannot quickly scale to match the massive surge in AI training and inference demand.
Get the full analysis with uListen AI
GPU scarcity is creating openings for new clouds and alternative AI hardware players.
Companies like CoreWeave, FoundryML, Cerebras, and Groq are seeing strong pull as customers seek non-traditional GPU access and are more willing to adopt specialized AI chips and federated GPU clouds.
Get the full analysis with uListen AI
Compute efficiency research will gain value when scaling is hardware-constrained.
Techniques like model distillation, smarter data selection, dynamic routing (e. ...
Get the full analysis with uListen AI
AI adoption is still in the earliest innings, especially for enterprises.
So far, mainly AI-native companies and a first wave of startups and tech-forward incumbents have adopted LLMs; true large-scale enterprise deployments are likely one to two years away due to long planning and prototyping cycles.
Get the full analysis with uListen AI
Vertical, tightly scoped AI agents are more likely to succeed initially.
Rather than building vague “do everything” assistants, founders should target specific, concrete workflows (e. ...
Get the full analysis with uListen AI
Agent infrastructure can be a business, but timing and position matter.
Some infrastructure layers succeed standalone (like Stripe), while others emerge from successful products with user scale (like Facebook auth), so builders must judge whether to start with tooling or with a vertical product that later exposes its platform.
Get the full analysis with uListen AI
Many 2021-era unicorns face a reckoning as cash burns and growth lags.
Elad predicts roughly a third will fail, a third have already hit peak valuation, and only a third will grow past it; founders should reset expectations, adjust burn, and focus on real revenue and product–market fit rather than preserving paper valuations.
Get the full analysis with uListen AI
Notable Quotes
“It's as if half the companies in the world over a year-long period decided, 'Yeah, we need supercomputers.'”
— Elad Gil
“I think we're in inning one.”
— Sarah Guo
“Usually starting with everything means you're not really doing anything deeply or well.”
— Elad Gil
“All I want to do is never write boilerplate code again.”
— Sarah Guo
“You’re really giving up the best years of your life working on things that potentially may not work.”
— Elad Gil
Questions Answered in This Episode
How long do you realistically expect the current GPU crunch to last, and what specific milestones would indicate that supply is finally catching up with AI demand?
Sarah Guo and Elad Gil discuss the current GPU shortage, its causes in semiconductor supply chains, and the surge in AI-driven demand that outpaces manufacturing capacity. ...
Get the full analysis with uListen AI
For founders building AI agents today, what are the most promising vertical workflows where you’d start, and which ones would you explicitly avoid?
Get the full analysis with uListen AI
How should startups evaluate whether to adopt alternative AI hardware (like Cerebras or Groq) versus waiting in line for NVIDIA GPUs?
Get the full analysis with uListen AI
What concrete metrics or thresholds should 2021-era startups use to decide whether to radically cut burn, pivot their business, or wind down?
Get the full analysis with uListen AI
In a world of constrained compute, which efficiency techniques (distillation, data curation, routing, etc.) do you believe will produce the biggest improvements in real-world AI applications?
Get the full analysis with uListen AI
Transcript Preview
Hey everyone. Welcome to No Priors. I'm Sarah Gua.
I'm Elad Gal.
This week on No Priors we're back with another episode where we answer your questions about tech, AI and everything in between.
I think we have a lot of different questions that people have brought up this week that they were hoping we could cover and some topics that we thought were- would be kind of interesting.
I wanna go to one of our listener questions and I think a topic that's really popular with many of the companies that you and I work with i- in terms of access to computing for much smaller scale experiments.
What, uh, what's going on with the GPU crunch?
Yeah. The companies that you and I work with, many of them are, uh, companies that, you know, they need to use very specific infrastructure to train and serve large models, right? Um, these work on GPUs and the structure of the industry is like, um, it's just not very robust, right? So you have a very small number of producers, NVIDIA and AMB- AMD generally, and then NVIDIA is very far ahead on the high-end processors that are most efficient for large-scale training and inference. Uh, then you have the pandemic supply disruption which we haven't fully recovered for. If you actually look at the supply chain, um, you go from the actual designers to, you know, the reliance on a few major foundries like TSMC. Um, you know, expansion of this capacity is not easy, right? New fabs are billions of dollars. Yield is a very complicated thing. You can think of it as a massive precision manufacturing problem where temperature, pressure, chemical concentration, tool imperfections, new processes, materials issues, like anything can make production have lower yield or lower quality, right? And- and so like if you think about the speed with which, uh, the industry, um, driven by both large and small players has decided that they want to do AI, uh, like the physical processes cannot keep up with that demand. It's as if, you know, half the companies in the world over a year-long period decided like, "Yeah, we need super computers." Not superconductors but, uh, gigantic networked GPUs.
So like what is the actual gap? So you're- to your point it sounds like, uh, much of the AI world is dependent on GPUs in order to train and then, uh, do inference on these big AI models and the big suppliers are basically NVIDIA, AMD and then there's like a long tail of smaller folks. What is the delta between the amount of capacity that exists today and that's needed? Are we off by 2X, 10X, some other number?
It's hard to say because right now there's no way to explore like the price elasticity of these things, right? Um, so, you know, just very specifically like the industry is kind of looking at deliveries in small quantity in September, larger quantities in December, January. Most of the large cloud providers are sold out for any scale for at least through April of next year. And so you have like really interesting dynamics like large cloud players who, you know, are the biggest consumers of- of these GPUs already, uh, like a Microsoft going and buying from other providers for near-term- near-term supply, right? So I think one question that I ask you is like, "Hey do you think this is a long-term thing? Do you think it's a very short-term thing?" But I- I think it just goes back to like the- the fundamental dynamics are do you expect the demand for these chips to continue increasing at a pace that out-increases the ability to scale a very physical like real-world process, right? Just to even be more specific, one of the challenges, like I was- I was talking to Jensen about this and a bonder, like not part of the GPU itself but like a critical tool in the manufacturing and assembly of GPUs is very specialized and so the ability to build any of these tools as well to enable these processes is- is a blocker. If you look at the demand from large labs today to continue increasing model scale and training time by magnitudes, I think it's hard to see that dynamic going away. What do you think?
Install uListen to search the full transcript and get AI-powered insights
Get Full TranscriptGet more from every podcast
AI summaries, searchable transcripts, and fact-checking. Free forever.
Add to Chrome