Skip to content
Dwarkesh PodcastDwarkesh Podcast

Ilya Sutskever (OpenAI Chief Scientist) — Why next-token prediction could surpass human intelligence

Asked Ilya Sutskever (Chief Scientist of OpenAI) about: * time to AGI * leaks and spies * what's after generative models * post AGI futures * working with MSFT and competing with Google * difficulty of aligning superhuman AI Hope you enjoy as much as I did! 𝐄𝐏𝐈𝐒𝐎𝐃𝐄 𝐋𝐈𝐍𝐊𝐒 * Transcript: https://www.dwarkeshpatel.com/p/ilya-sutskever * Apple Podcasts: https://apple.co/42H6c4D * Spotify: https://spoti.fi/3LRqOBd * Follow me on Twitter: https://twitter.com/dwarkesh_sp 𝐓𝐈𝐌𝐄𝐒𝐓𝐀𝐌𝐏𝐒 00:00:00 - Time to AGI 00:05:57 - What’s after generative models? 00:10:57 - Data, models, and research 00:15:27 - Alignment 00:20:53 - Post AGI Future 00:26:56 - New ideas are overrated 00:36:22 - Is progress inevitable? 00:41:27 - Future Breakthroughs

Ilya SutskeverguestDwarkesh Patelhost
Mar 26, 202347mWatch on YouTube ↗

At a glance

WHAT IT’S REALLY ABOUT

Ilya Sutskever defends next‑token prediction as path to superintelligence

  1. Ilya Sutskever, OpenAI’s cofounder and chief scientist, argues that next‑token prediction, when scaled sufficiently, can both match and surpass human intelligence by implicitly modeling the underlying reality and human cognition that generate language.
  2. He discusses the likely economic and societal trajectory from powerful AI to AGI, emphasizing reliability and controllability as the key bottlenecks and most important emergent properties to aim for.
  3. Sutskever outlines how current systems are trained (RLHF, human‑in‑the‑loop, AI‑generated data), explains why OpenAI pivoted away from robotics, and describes how alignment will require multiple, complementary approaches rather than a single mathematical definition.
  4. He reflects on AGI’s long‑term societal impact, the inevitability of AI progress given hardware and data trends, and the role future AIs will play in both AI research and human inner development.

IDEAS WORTH REMEMBERING

5 ideas

Next‑token prediction can, in principle, exceed human performance.

Sutskever argues that if a base model learns to predict text extremely well, it must internalize the structure of reality and human minds, allowing it to extrapolate the behavior of hypothetical agents smarter than any real human.

Reliability is the central bottleneck to AI’s economic impact.

He notes that if 2030 arrives with disappointing AI‑driven GDP gains, the likely culprit will be models that still require extensive human checking, limiting automation and trust in high‑stakes domains.

Alignment will require multiple overlapping methods, not one formula.

Instead of a single clean mathematical definition of alignment, he anticipates a toolbox: adversarial testing, behavioral probes, interpretability tools, and smaller “verifier” models inspecting larger ones.

Human‑AI collaboration in training is preferable to fully autonomous self‑improvement.

He envisions a regime where humans provide a small fraction of high‑quality signals while AIs generate most training data, preserving human oversight rather than moving to 100% AI‑only feedback.

Data and hardware are still enabling scale, but new training routes are needed.

While text data has not yet run out and GPUs remain adequate, Sutskever expects that future progress will increasingly rely on synthetic data, multimodal inputs, and algorithmic improvements as natural data becomes scarce.

WORDS WORTH SAVING

5 quotes

I challenge the claim that next token prediction cannot surpass human performance.

Ilya Sutskever

Predicting the next token well means that you understand the underlying reality that led to the creation of that token.

Ilya Sutskever

I would not underestimate the difficulty of alignment of models that are actually smarter than us, of models that are capable of misrepresenting their intentions.

Ilya Sutskever

The main activity is actually understanding… it was a new understanding of very old things.

Ilya Sutskever

Change is the only constant… I don’t think anyone has any idea of how the world will look like in 3000.

Ilya Sutskever

Next‑token prediction and whether it can surpass human intelligenceEconomic impact of AI, timelines to AGI, and model commoditizationReliability, controllability, and alignment strategies for advanced modelsTraining paradigms: RLHF, AI‑generated data, scaling laws, and data limitsResearch directions: multimodality, robotics, architectures, and algorithmic advancesHardware, data, and the inevitability/contingency of AI progressSocietal futures with AGI: meaning, human–AI integration, and governance

High quality AI-generated summary created from speaker-labeled transcript.

Get more out of YouTube videos.

High quality summaries for YouTube videos. Accurate transcripts to search & find moments. Powered by ChatGPT & Claude AI.

Add to Chrome