Dwarkesh PodcastIlya Sutskever (OpenAI Chief Scientist) — Why next-token prediction could surpass human intelligence
At a glance
WHAT IT’S REALLY ABOUT
Ilya Sutskever defends next‑token prediction as path to superintelligence
- Ilya Sutskever, OpenAI’s cofounder and chief scientist, argues that next‑token prediction, when scaled sufficiently, can both match and surpass human intelligence by implicitly modeling the underlying reality and human cognition that generate language.
- He discusses the likely economic and societal trajectory from powerful AI to AGI, emphasizing reliability and controllability as the key bottlenecks and most important emergent properties to aim for.
- Sutskever outlines how current systems are trained (RLHF, human‑in‑the‑loop, AI‑generated data), explains why OpenAI pivoted away from robotics, and describes how alignment will require multiple, complementary approaches rather than a single mathematical definition.
- He reflects on AGI’s long‑term societal impact, the inevitability of AI progress given hardware and data trends, and the role future AIs will play in both AI research and human inner development.
IDEAS WORTH REMEMBERING
5 ideasNext‑token prediction can, in principle, exceed human performance.
Sutskever argues that if a base model learns to predict text extremely well, it must internalize the structure of reality and human minds, allowing it to extrapolate the behavior of hypothetical agents smarter than any real human.
Reliability is the central bottleneck to AI’s economic impact.
He notes that if 2030 arrives with disappointing AI‑driven GDP gains, the likely culprit will be models that still require extensive human checking, limiting automation and trust in high‑stakes domains.
Alignment will require multiple overlapping methods, not one formula.
Instead of a single clean mathematical definition of alignment, he anticipates a toolbox: adversarial testing, behavioral probes, interpretability tools, and smaller “verifier” models inspecting larger ones.
Human‑AI collaboration in training is preferable to fully autonomous self‑improvement.
He envisions a regime where humans provide a small fraction of high‑quality signals while AIs generate most training data, preserving human oversight rather than moving to 100% AI‑only feedback.
Data and hardware are still enabling scale, but new training routes are needed.
While text data has not yet run out and GPUs remain adequate, Sutskever expects that future progress will increasingly rely on synthetic data, multimodal inputs, and algorithmic improvements as natural data becomes scarce.
WORDS WORTH SAVING
5 quotesI challenge the claim that next token prediction cannot surpass human performance.
— Ilya Sutskever
Predicting the next token well means that you understand the underlying reality that led to the creation of that token.
— Ilya Sutskever
I would not underestimate the difficulty of alignment of models that are actually smarter than us, of models that are capable of misrepresenting their intentions.
— Ilya Sutskever
The main activity is actually understanding… it was a new understanding of very old things.
— Ilya Sutskever
Change is the only constant… I don’t think anyone has any idea of how the world will look like in 3000.
— Ilya Sutskever
High quality AI-generated summary created from speaker-labeled transcript.
Get more out of YouTube videos.
High quality summaries for YouTube videos. Accurate transcripts to search & find moments. Powered by ChatGPT & Claude AI.
Add to Chrome