Dwarkesh PodcastDemis Hassabis — Scaling, superhuman AIs, AlphaZero atop LLMs, AlphaFold
At a glance
WHAT IT’S REALLY ABOUT
Demis Hassabis explains path to AGI, scaling, safety, and Gemini
- Demis Hassabis outlines how large multimodal models plus planning/search (AlphaZero-style RL) are his likely recipe for AGI within the next decade, with current LLMs seen as “unreasonably effective” but still incomplete.
- He emphasizes neuroscience-inspired ideas like world models, imagination, and sample-efficient learning, and argues that combining old deep RL advances with new large-model scaling will be crucial.
- A large portion of the conversation covers safety: grounding, deception, evals, sandboxes, cybersecurity, responsible scaling, and governance that involves governments, academia, and civil society rather than just private firms.
- Hassabis also discusses Gemini’s development, multimodality, robotics, AI-for-science applications like AlphaFold, and how AGI and proto-AGI systems could accelerate future AI research while requiring tight control and oversight.
IDEAS WORTH REMEMBERING
5 ideasAGI will likely require both scaled multimodal models and explicit planning/search.
Hassabis expects future AGI systems to use large multimodal world models (like Gemini) as priors, with AlphaZero-style planning and search on top to perform deliberate reasoning and goal-directed behavior that current LLMs lack.
Neuroscience remains a powerful guide for AI, especially around world models and imagination.
Concepts like reinforcement learning, attention, experience replay, and mental simulation were inspired by neuroscience and will inform unresolved areas such as planning, world-model construction, and imagination-like mental simulation in machines.
Scaling has gone further than expected, but its ultimate limits are still empirical.
Large models show emergent abstractions and partial grounding from language alone, which even scaling proponents didn’t fully anticipate; Hassabis argues we must push scaling and innovation in parallel to see whether we hit a soft asymptote or a “brick wall.”
Synthetic data and self-play will be central to overcoming data bottlenecks.
DeepMind plans to leverage realistic simulations, self-play between agents, and targeted synthetic data generation to fill gaps in training distributions, using careful data analysis to identify underrepresented regions and reduce bias.
Robust safety requires evals for deception and misuse, secure infrastructure, and sandboxes.
Hassabis stresses the need for better benchmarks (e.g., for deception, code exfiltration), hardened sandbox environments, strong cybersecurity, and even narrow AIs to help inspect more capable “general” systems before wider deployment.
WORDS WORTH SAVING
5 quotesI wouldn’t be surprised if we had AGI-like systems within the next decade.
— Demis Hassabis
I sort of look at the large models today and I think they’re almost unreasonably effective for what they are.
— Demis Hassabis
The brain is an existence proof that general intelligence is possible at all.
— Demis Hassabis
You don’t want to be live A/B testing out in the world with these very consequential systems.
— Demis Hassabis
This is so consequential, this technology, I think it’s much bigger than any one company or even industry in general.
— Demis Hassabis
High quality AI-generated summary created from speaker-labeled transcript.
Get more out of YouTube videos.
High quality summaries for YouTube videos. Accurate transcripts to search & find moments. Powered by ChatGPT & Claude AI.
Add to Chrome