Y CombinatorWhy The Next AI Breakthroughs Will Be In Reasoning, Not Scaling
At a glance
WHAT IT’S REALLY ABOUT
AI’s Next Leap: Reasoning Engines Transform Science, Chips, And Startups
- The hosts discuss Sam Altman’s recent AGI/ASI essay and argue that the next major AI breakthroughs will come from reasoning-focused models like OpenAI’s o1, not just larger model scaling. They highlight how o1’s chain-of-thought and reinforcement-learning-inspired architecture unlocks step-function improvements in hard domains such as chip design, CAD/airfoils, and complex customer support. Several YC-backed companies are showcased as concrete examples of o1 enabling system-level engineering, advanced physics simulations, and high-accuracy automation that older models like GPT-4o couldn’t reliably handle. The conversation closes with implications for startup moats, the centrality of evals and proprietary data, and how reasoning models may usher in a “Star Trek” style age of physical-world abundance if steered well.
IDEAS WORTH REMEMBERING
5 ideasReasoning-focused models mark a step-change beyond simple scaling.
OpenAI’s o1 introduces chain-of-thought and reinforcement-learning-style training that lets models ‘think through’ problems, enabling capabilities (like complex chip system design) that GPT-4o could not handle with the same prompts.
AI can now meaningfully automate expert-level hardware and engineering tasks.
Companies like Diode Computer and Camfer use o1 to perform high-level PCB system design, component selection, and multi-equation airfoil simulations—work that previously required specialized electrical or mechanical engineers.
Eval sets and proprietary workflows are becoming core startup moats.
The hosts argue that writing tens of thousands of high-quality eval cases, especially using non-public, domain-specific data, is a durable advantage when everyone accesses similar base models.
Advanced reasoning dramatically increases automation viability in messy domains.
Gigaml’s customer support product jumped from about 70% error in hard cases to roughly 5% error (around 85% accuracy) using o1 and rigorous evals, making AI agents credible for complex, non-rules-based support.
Strong technical teams matter more, not less, in the o1 era.
While some fear AI will commoditize engineering, the panel believes the highest value will accrue to teams that can push models the final 10%—through clever prompts, evals, UI, integrations, and domain-specific reasoning.
WORDS WORTH SAVING
5 quotesIt’s the worst that these models are ever going to be right now, right this moment.
— Gary
What was missing from its ability to actually do science and accelerate technological progress is it needs to be able to think through things.
— Jared
They went from 0% accuracy to 85% accuracy.
— Diana (about Gigaml’s o1-powered customer support)
All of the value is probably going to be captured by the strongest technical teams who can build on top of whatever the base level of tech is and get the final 10%.
— Harj
It can’t just be helping people click a little bit faster. It’s gotta be things that actually create real-world abundance for everyone.
— Gary
High quality AI-generated summary created from speaker-labeled transcript.
Get more out of YouTube videos.
High quality summaries for YouTube videos. Accurate transcripts to search & find moments. Powered by ChatGPT & Claude AI.
Add to Chrome