No PriorsNo Priors Ep. 7 | With Stanford Professor Dr. Percy Liang
At a glance
WHAT IT’S REALLY ABOUT
Percy Liang on Foundation Models, Open Research, and AI’s Future Impact
- Stanford professor Percy Liang discusses his journey in natural language processing and how GPT-3 triggered his shift toward foundation models, leading to the creation of Stanford’s Center for Research on Foundation Models (CRFM).
- He explains why large language models are a paradigm shift, their emergent behaviors like in-context learning and chain-of-thought, and how academia’s role is moving from “making things work” to understanding principles and social impact.
- Liang highlights CRFM’s work on transparency (e.g., the HELM evaluation benchmark), interdisciplinary research on risks and benefits, and efforts to keep models accessible despite increasing industry secrecy.
- He also covers Together, a decentralized compute initiative for open models, and reflects on future directions for architectures beyond transformers, scientific discovery with AI, and how to think rigorously about AGI.
IDEAS WORTH REMEMBERING
5 ideasGPT-3’s training paradigm transformed how we conceptualize AI tasks.
Simply predicting the next word at scale, and then prompting models in natural language, turns a single foundation model into a flexible substrate for many tasks, dissolving rigid, task-specific system design.
Academia’s comparative advantage is shifting from scaling models to understanding them.
With industry able to “just scale” using massive data and compute, universities are better positioned to study principles, data/architecture effects, robustness, and social impacts rather than only chasing benchmark wins.
Transparency and openness are eroding and must be deliberately rebuilt.
Unlike earlier deep learning culture (open datasets, code, and models), top foundation models are now often API-only; CRFM and HELM aim to reintroduce norms of disclosure, comparability, and shared benchmarks.
Emergent behaviors like in-context learning and chain-of-thought weren’t hand-designed.
Capabilities such as learning from prompts, step-by-step reasoning, and stylistic “mix and match” arise from scale and training rather than explicit programming, suggesting deeper, still poorly understood structure.
Future AI progress likely requires both new architectures and better infrastructure.
Liang expects transformers not to be the final architecture and supports work on alternative models, while Together tackles the compute bottleneck via decentralized, weaker interconnects to make large-scale training more accessible.
WORDS WORTH SAVING
5 quotesThe idea of a task, which is so central to AI, begins to dissolve.
— Percy Liang
I really hope that in 10 years we won’t be using the transformer.
— Percy Liang
We coined the term foundation models because ‘large language models’ didn’t really capture the significance.
— Percy Liang
Up until now, the AI dream tops out at humans, but now we can actually go beyond in many, many ways.
— Percy Liang
Transparency is necessary but not sufficient; you need it just to even have a conversation about policy.
— Percy Liang
High quality AI-generated summary created from speaker-labeled transcript.
Get more out of YouTube videos.
High quality summaries for YouTube videos. Accurate transcripts to search & find moments. Powered by ChatGPT & Claude AI.
Add to Chrome