
Nvidia Part III: The Dawn of the AI Era (2022-2023) (Audio)
Ben Gilbert (host), David Rosenthal (host)
In this episode of Acquired, featuring Ben Gilbert and David Rosenthal, Nvidia Part III: The Dawn of the AI Era (2022-2023) (Audio) explores how Nvidia’s data-center platform powered the generative AI boom Ben Gilbert and David Rosenthal revisit Nvidia because the 2022–2023 generative AI breakout (ChatGPT and LLMs) created a step-function shift in data-center demand for GPU compute.
How Nvidia’s data-center platform powered the generative AI boom
Ben Gilbert and David Rosenthal revisit Nvidia because the 2022–2023 generative AI breakout (ChatGPT and LLMs) created a step-function shift in data-center demand for GPU compute.
They trace the technical lineage from AlexNet (2012) to transformers (2017), then to OpenAI’s pivot to for-profit funding to afford massive training runs—culminating in ChatGPT’s “Netscape/iPhone moment.”
The episode argues Nvidia was uniquely prepared: CUDA’s developer ecosystem, data-center “full stack” systems (DGX, Grace CPU, Hopper GPUs), and the Mellanox/InfiniBand networking acquisition enabled multi-GPU, multi-rack training at scale.
They close with business analysis: Nvidia’s platform-like moats (software, switching costs, constrained advanced packaging capacity) and the key uncertainty—whether AI value creation is durable enough to justify the current capital spending wave.
Key Takeaways
Transformers made language modeling massively parallel—and GPUs made it economically feasible.
Attention allows models to use wide context, but it’s computationally heavy (quadratic with context length). ...
Get the full analysis with uListen AI
OpenAI’s success depended as much on capital structure and cloud access as on research.
To compete with Google’s resources, OpenAI created a capped-profit entity (2019) and partnered with Microsoft for funding and exclusive cloud compute—enabling GPT-3 through ChatGPT and GPT-4.
Get the full analysis with uListen AI
Nvidia’s advantage is “platform + system,” not just a fast chip.
CUDA (compiler/runtime/libraries) plus integrated hardware (DGX, NVLink, Grace CPU) and networking (InfiniBand) creates an end-to-end solution that’s hard to replicate with piecemeal components.
Get the full analysis with uListen AI
The bottleneck in modern AI shifted from compute to memory and interconnect.
LLMs require huge on-package HBM and extremely fast chip-to-chip and rack-to-rack communication. ...
Get the full analysis with uListen AI
Mellanox/InfiniBand became a strategic masterstroke once “the data center is the computer.”
Training frontier models requires treating hundreds/thousands of GPUs as one machine. ...
Get the full analysis with uListen AI
Nvidia is monetizing scarcity and integration with extraordinary margins.
H100 pricing (~$40k per GPU) and DGX systems (hundreds of thousands to “call us” scale) plus forecast ~70%+ gross margins show pricing power driven by demand exceeding CoWoS/HBM supply.
Get the full analysis with uListen AI
The key risk isn’t “AMD vs Nvidia,” but disintermediation and shifting compute mix.
Hyperscalers building custom silicon, frameworks (e. ...
Get the full analysis with uListen AI
Notable Quotes
“In our April 2022 episodes, we never once said the word 'generative.' That is how fast things have changed.”
— Ben Gilbert
“In November of 2022, AI definitely had its Netscape moment… it may have even been its iPhone moment.”
— Ben Gilbert
“The more accurately an LLM predicts that next word… ipso facto, the greater its understanding.”
— David Rosenthal (summarizing Ilya Sutskever)
“The data center is the computer.”
— Ben Gilbert (describing Jensen Huang’s framing)
“Starting price for DGX Cloud is $37,000 a month… that’s three-month payback on the CapEx.”
— David Rosenthal
Questions Answered in This Episode
What specific technical constraints (HBM capacity, NVLink, InfiniBand bandwidth) most limit frontier-model training today, and which is likely to ease first?
Ben Gilbert and David Rosenthal revisit Nvidia because the 2022–2023 generative AI breakout (ChatGPT and LLMs) created a step-function shift in data-center demand for GPU compute.
Get the full analysis with uListen AI
How decisive was OpenAI’s 2019 for-profit pivot versus the 2017 transformer invention in making ChatGPT inevitable?
They trace the technical lineage from AlexNet (2012) to transformers (2017), then to OpenAI’s pivot to for-profit funding to afford massive training runs—culminating in ChatGPT’s “Netscape/iPhone moment.”
Get the full analysis with uListen AI
To what extent is CUDA a true lock-in today—how much of the moat is software ecosystem versus hardware supply (CoWoS/HBM capacity)?
The episode argues Nvidia was uniquely prepared: CUDA’s developer ecosystem, data-center “full stack” systems (DGX, Grace CPU, Hopper GPUs), and the Mellanox/InfiniBand networking acquisition enabled multi-GPU, multi-rack training at scale.
Get the full analysis with uListen AI
If inference becomes the dominant spend, what parts of Nvidia’s stack (CUDA libraries, DGX Cloud, networking) still provide durable differentiation?
They close with business analysis: Nvidia’s platform-like moats (software, switching costs, constrained advanced packaging capacity) and the key uncertainty—whether AI value creation is durable enough to justify the current capital spending wave.
Get the full analysis with uListen AI
How credible is Jensen’s reframed “$1T data-center TAM” argument, and what would have to be true about enterprise AI ROI for it to sustain?
Get the full analysis with uListen AI
Transcript Preview
You like my Buck's T-shirt?
I love your Buck's T-shirt.
I went for the first time, what, two weeks ago, when I was down for a meeting at Benchmark, and the nostalgia in there was just unbelievable.
I can't believe you hadn't been before. I know Jensen is a Denny's guy, but I feel like he would meet us at Buck's if we asked him.
Or at the very least, we should figure out some Nvidia memorabilia to get on the wall at Buck's.
Totally.
Fit right in. All right, let's do it.
Let's do it.
Who got the truth? Is it you? Is it you? Is it you? Who got the truth now? Is it you? Is it you? Is it you? Sit me down, say it straight. Another story on the way. Who got the truth?
Welcome to Season 13, Episode 3 of Acquired, the podcast about great technology companies and the stories and playbooks behind them. I'm Ben Gilbert.
I'm David Rosenthal.
And we are your hosts. Today, we tell a story that we thought we had already finished: Nvidia. But the last eighteen months have been so insane, listeners, that it warranted an entire episode on its own. So today is a part three for us with Nvidia, telling the story of the AI revolution, how we got here, and why it's happening now, starting all the way down at the level of atoms and silicon. So here's something crazy that I did a transcript search on to see if it was true. In our April 2022 episodes, we never once said the word "generative." That is how fast things have changed.
Unbelievable.
Totally crazy. And the timing of all of this AI stuff in the world is unbelievably coincidental and, uh, very favorable. So recall back to eighteen months ago. Throughout 2022, we all watched financial markets, from public equities to early-stage startups to real estate, just fall off a cliff due to rapid rise in interest rates. The crypto and Web3 bubble burst, banks fail. It seemed like the whole tech economy, and potentially a lot with it, was heading into a long winter.
Including Nvidia.
Including Nvidia, who had that massive inventory write-off for what they thought was over-ordering.
Yep. Wow, how things have changed.
[chuckles] Yeah. But by the fall of 2022, right when everything looked the absolute bleakest, a breakthrough technology finally became useful after years in research labs. Large language models, or LLMs, built on the innovative transformer machine learning mechanism, burst onto the scene, first with OpenAI's ChatGPT, which became the fastest app in history to a hundred million active users, and then quickly followed by Microsoft, Google, and seemingly every other company. In November of 2022, AI definitely had its Netscape moment, and time will tell, but it may have even been its iPhone moment.
Well, that is definitely what Jensen believes.
Yep. Well, today, we'll explore exactly how this breakthrough came to be, the individuals behind it, and of course, why the entire thing has happened on top of Nvidia's hardware and software. If you wanna make sure you know every time there's a new episode, go sign up at acquired.fm/email. You'll also get access to two things that we aren't putting anywhere else: one, a clue as to what the next episode will be, and two, follow-ups from previous episodes from things that we learned after release. You can come talk about this episode with us after listening at acquired.fm/slack. If you want more of David and I, check out our interview show, ACQ2. Our next few episodes are about AI, with CEOs leading the way in this world we are talking about today, and a great interview with Doug Demuro, where, uh, we wanted to talk about a lot more than just Porsche with him, but, uh, you know, we only had eleven hours or whatever we had-
Install uListen to search the full transcript and get AI-powered insights
Get Full TranscriptGet more from every podcast
AI summaries, searchable transcripts, and fact-checking. Free forever.
Add to Chrome