Nvidia Part III: The Dawn of the AI Era (2022-2023) (Audio)

Nvidia Part III: The Dawn of the AI Era (2022-2023) (Audio)

AcquiredSep 6, 20232h 54m

Ben Gilbert (host), David Rosenthal (host)

AlexNet and the 2012 AI “Big Bang”OpenAI’s founding motivation and Microsoft partnershipTransformers and the attention mechanism (parallelism, O(n^2))Von Neumann bottleneck and GPU accelerationScaling laws: parameters, data, and emergent capabilitiesNvidia’s data-center stack: Hopper, Grace, DGX, NVLinkMellanox/InfiniBand and AI cluster networkingTSMC CoWoS + HBM packaging constraintsDGX Cloud and shifting customer relationshipsNvidia moats: CUDA ecosystem, switching costs, cornered capacity2023 earnings explosion and the reframed “$1T TAM”Bull vs bear: durability of AI demand and competitive responses

In this episode of Acquired, featuring Ben Gilbert and David Rosenthal, Nvidia Part III: The Dawn of the AI Era (2022-2023) (Audio) explores how Nvidia’s data-center platform powered the generative AI boom Ben Gilbert and David Rosenthal revisit Nvidia because the 2022–2023 generative AI breakout (ChatGPT and LLMs) created a step-function shift in data-center demand for GPU compute.

How Nvidia’s data-center platform powered the generative AI boom

Ben Gilbert and David Rosenthal revisit Nvidia because the 2022–2023 generative AI breakout (ChatGPT and LLMs) created a step-function shift in data-center demand for GPU compute.

They trace the technical lineage from AlexNet (2012) to transformers (2017), then to OpenAI’s pivot to for-profit funding to afford massive training runs—culminating in ChatGPT’s “Netscape/iPhone moment.”

The episode argues Nvidia was uniquely prepared: CUDA’s developer ecosystem, data-center “full stack” systems (DGX, Grace CPU, Hopper GPUs), and the Mellanox/InfiniBand networking acquisition enabled multi-GPU, multi-rack training at scale.

They close with business analysis: Nvidia’s platform-like moats (software, switching costs, constrained advanced packaging capacity) and the key uncertainty—whether AI value creation is durable enough to justify the current capital spending wave.

Key Takeaways

Transformers made language modeling massively parallel—and GPUs made it economically feasible.

Attention allows models to use wide context, but it’s computationally heavy (quadratic with context length). ...

Get the full analysis with uListen AI

OpenAI’s success depended as much on capital structure and cloud access as on research.

To compete with Google’s resources, OpenAI created a capped-profit entity (2019) and partnered with Microsoft for funding and exclusive cloud compute—enabling GPT-3 through ChatGPT and GPT-4.

Get the full analysis with uListen AI

Nvidia’s advantage is “platform + system,” not just a fast chip.

CUDA (compiler/runtime/libraries) plus integrated hardware (DGX, NVLink, Grace CPU) and networking (InfiniBand) creates an end-to-end solution that’s hard to replicate with piecemeal components.

Get the full analysis with uListen AI

The bottleneck in modern AI shifted from compute to memory and interconnect.

LLMs require huge on-package HBM and extremely fast chip-to-chip and rack-to-rack communication. ...

Get the full analysis with uListen AI

Mellanox/InfiniBand became a strategic masterstroke once “the data center is the computer.”

Training frontier models requires treating hundreds/thousands of GPUs as one machine. ...

Get the full analysis with uListen AI

Nvidia is monetizing scarcity and integration with extraordinary margins.

H100 pricing (~$40k per GPU) and DGX systems (hundreds of thousands to “call us” scale) plus forecast ~70%+ gross margins show pricing power driven by demand exceeding CoWoS/HBM supply.

Get the full analysis with uListen AI

The key risk isn’t “AMD vs Nvidia,” but disintermediation and shifting compute mix.

Hyperscalers building custom silicon, frameworks (e. ...

Get the full analysis with uListen AI

Notable Quotes

In our April 2022 episodes, we never once said the word 'generative.' That is how fast things have changed.

Ben Gilbert

In November of 2022, AI definitely had its Netscape moment… it may have even been its iPhone moment.

Ben Gilbert

The more accurately an LLM predicts that next word… ipso facto, the greater its understanding.

David Rosenthal (summarizing Ilya Sutskever)

The data center is the computer.

Ben Gilbert (describing Jensen Huang’s framing)

Starting price for DGX Cloud is $37,000 a month… that’s three-month payback on the CapEx.

David Rosenthal

Questions Answered in This Episode

What specific technical constraints (HBM capacity, NVLink, InfiniBand bandwidth) most limit frontier-model training today, and which is likely to ease first?

Ben Gilbert and David Rosenthal revisit Nvidia because the 2022–2023 generative AI breakout (ChatGPT and LLMs) created a step-function shift in data-center demand for GPU compute.

Get the full analysis with uListen AI

How decisive was OpenAI’s 2019 for-profit pivot versus the 2017 transformer invention in making ChatGPT inevitable?

They trace the technical lineage from AlexNet (2012) to transformers (2017), then to OpenAI’s pivot to for-profit funding to afford massive training runs—culminating in ChatGPT’s “Netscape/iPhone moment.”

Get the full analysis with uListen AI

To what extent is CUDA a true lock-in today—how much of the moat is software ecosystem versus hardware supply (CoWoS/HBM capacity)?

The episode argues Nvidia was uniquely prepared: CUDA’s developer ecosystem, data-center “full stack” systems (DGX, Grace CPU, Hopper GPUs), and the Mellanox/InfiniBand networking acquisition enabled multi-GPU, multi-rack training at scale.

Get the full analysis with uListen AI

If inference becomes the dominant spend, what parts of Nvidia’s stack (CUDA libraries, DGX Cloud, networking) still provide durable differentiation?

They close with business analysis: Nvidia’s platform-like moats (software, switching costs, constrained advanced packaging capacity) and the key uncertainty—whether AI value creation is durable enough to justify the current capital spending wave.

Get the full analysis with uListen AI

How credible is Jensen’s reframed “$1T data-center TAM” argument, and what would have to be true about enterprise AI ROI for it to sustain?

Get the full analysis with uListen AI

Transcript Preview

Ben Gilbert

You like my Buck's T-shirt?

David Rosenthal

I love your Buck's T-shirt.

Ben Gilbert

I went for the first time, what, two weeks ago, when I was down for a meeting at Benchmark, and the nostalgia in there was just unbelievable.

David Rosenthal

I can't believe you hadn't been before. I know Jensen is a Denny's guy, but I feel like he would meet us at Buck's if we asked him.

Ben Gilbert

Or at the very least, we should figure out some Nvidia memorabilia to get on the wall at Buck's.

David Rosenthal

Totally.

Ben Gilbert

Fit right in. All right, let's do it.

David Rosenthal

Let's do it.

Speaker

Who got the truth? Is it you? Is it you? Is it you? Who got the truth now? Is it you? Is it you? Is it you? Sit me down, say it straight. Another story on the way. Who got the truth?

Ben Gilbert

Welcome to Season 13, Episode 3 of Acquired, the podcast about great technology companies and the stories and playbooks behind them. I'm Ben Gilbert.

David Rosenthal

I'm David Rosenthal.

Ben Gilbert

And we are your hosts. Today, we tell a story that we thought we had already finished: Nvidia. But the last eighteen months have been so insane, listeners, that it warranted an entire episode on its own. So today is a part three for us with Nvidia, telling the story of the AI revolution, how we got here, and why it's happening now, starting all the way down at the level of atoms and silicon. So here's something crazy that I did a transcript search on to see if it was true. In our April 2022 episodes, we never once said the word "generative." That is how fast things have changed.

David Rosenthal

Unbelievable.

Ben Gilbert

Totally crazy. And the timing of all of this AI stuff in the world is unbelievably coincidental and, uh, very favorable. So recall back to eighteen months ago. Throughout 2022, we all watched financial markets, from public equities to early-stage startups to real estate, just fall off a cliff due to rapid rise in interest rates. The crypto and Web3 bubble burst, banks fail. It seemed like the whole tech economy, and potentially a lot with it, was heading into a long winter.

David Rosenthal

Including Nvidia.

Ben Gilbert

Including Nvidia, who had that massive inventory write-off for what they thought was over-ordering.

David Rosenthal

Yep. Wow, how things have changed.

Ben Gilbert

[chuckles] Yeah. But by the fall of 2022, right when everything looked the absolute bleakest, a breakthrough technology finally became useful after years in research labs. Large language models, or LLMs, built on the innovative transformer machine learning mechanism, burst onto the scene, first with OpenAI's ChatGPT, which became the fastest app in history to a hundred million active users, and then quickly followed by Microsoft, Google, and seemingly every other company. In November of 2022, AI definitely had its Netscape moment, and time will tell, but it may have even been its iPhone moment.

David Rosenthal

Well, that is definitely what Jensen believes.

Ben Gilbert

Yep. Well, today, we'll explore exactly how this breakthrough came to be, the individuals behind it, and of course, why the entire thing has happened on top of Nvidia's hardware and software. If you wanna make sure you know every time there's a new episode, go sign up at acquired.fm/email. You'll also get access to two things that we aren't putting anywhere else: one, a clue as to what the next episode will be, and two, follow-ups from previous episodes from things that we learned after release. You can come talk about this episode with us after listening at acquired.fm/slack. If you want more of David and I, check out our interview show, ACQ2. Our next few episodes are about AI, with CEOs leading the way in this world we are talking about today, and a great interview with Doug Demuro, where, uh, we wanted to talk about a lot more than just Porsche with him, but, uh, you know, we only had eleven hours or whatever we had-

Install uListen to search the full transcript and get AI-powered insights

Get Full Transcript

Get more from every podcast

AI summaries, searchable transcripts, and fact-checking. Free forever.

Add to Chrome