The Twenty Minute VCZico Kolter: OpenAI's Newest Board Member on The Biggest Questions and Concerns in AI Safety | E1197
At a glance
WHAT IT’S REALLY ABOUT
OpenAI Board Member Zico Kolter Dissects Data, Safety, and AGI Futures
- Zico Kolter, head of CMU’s Machine Learning Department and new OpenAI board member, explains why next-word prediction LLMs are a profound scientific discovery and why we are far from hitting hard limits on data, models, or compute.
- He argues that data is not the core bottleneck, model architectures are increasingly commoditized, and that larger models still deliver meaningful gains—especially in complex tasks like coding—despite benchmark plateaus.
- Kolter’s central AI safety concern is that current models cannot reliably follow specifications, making them vulnerable to prompt injection and jailbreaks, which becomes dangerous as we embed them into critical systems and agents.
- He favors a pragmatic focus on near-term risks like cyberattacks, misinformation, and infrastructure failures, is cautious but not absolutist on open-weight releases, and remains broadly optimistic that society can adapt if safety is treated as a prerequisite for deployment.
IDEAS WORTH REMEMBERING
5 ideasData is not the near-term bottleneck for AI progress.
Despite having used much of the highest-quality public text, Kolter notes that models are currently trained on surprisingly small datasets (tens of terabytes) relative to what exists, and vast untapped multimodal and private data—constrained more by compute and methods than sheer availability—remain.
Model architectures matter less than scale, data, and training strategy.
Kolter believes we are in a “post-architecture” phase: transformers are useful but not uniquely magical, and many architectures could work if scaled and trained similarly; capabilities are driven more by data, size, and optimization than clever structural tweaks.
Larger frontier models still provide meaningful real-world gains.
While benchmarks show diminishing improvements (e.g., 92% vs. 94%), Kolter sees substantial qualitative gains in tasks like coding and lecture processing, suggesting users underestimate what newer models can do rather than models having plateaued.
The most urgent safety problem is unreliable adherence to specifications.
Because models can be prompt-injected or jailbroken, they often override developer instructions in favor of user prompts; this is tolerable in chatbots but becomes critical when LLMs are embedded in agents and infrastructure, effectively creating an unpatchable ‘buffer overflow’ style vulnerability.
AI will drastically lower the skill bar for serious cyber and other attacks.
Kolter highlights cyber risk as especially acute: models that can find software vulnerabilities or craft exploits could put powerful attack capabilities into the hands of many low-skill actors, making even known risks far more scalable and dangerous.
WORDS WORTH SAVING
5 quotesYou can train word predictors and they produce intelligent, coherent, long-form responses; that is one of the most notable scientific discoveries of the past 10 or 20 years.
— Zico Kolter
We are nowhere close to hitting the limits of available data in these models.
— Zico Kolter
Right now the AI models we have are not able to reliably follow specifications.
— Zico Kolter
This is sort of like these models have a buffer overflow in all of them that we know about and that we don’t know how to patch and fix.
— Zico Kolter
I want to develop and improve safety of these tools because I want to use them. To reach that point, they have to be safe.
— Zico Kolter
High quality AI-generated summary created from speaker-labeled transcript.
Get more out of YouTube videos.
High quality summaries for YouTube videos. Accurate transcripts to search & find moments. Powered by ChatGPT & Claude AI.
Add to Chrome