Skip to content
No PriorsNo Priors

No Priors Ep. 118 | With Anthropic Co-Founder Ben Mann

What happens when you give AI researchers unlimited compute and tell them to compete for the highest usage rates? Ben Mann from Anthropic sits down with Sarah Guo and Elad Gil to explain how Claude 4 went from "reward hacking" to efficiently completing tasks and how they're racing to solve AI safety before deploying computer-controlling agents. Ben talks about economic Turing tests, the future of general versus specialized AI models, Reinforcement Learning From AI Feedback (RLAIF), and Anthropic’s Model Context Protocol (MCP). Plus, Ben shares his thoughts on if we will have Superintelligence by 2028. Sign up for new podcasts every week. Email feedback to show@no-priors.com Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil | @8enmann Links: ai-2027.com/ Chapters: 00:00 Ben Mann Introduction 00:33 Releasing Claude 4 02:05 Claude 4 Highlights and Improvements 03:42 Advanced Use Cases and Capabilities 06:42 Specialization and Future of AI Models 09:35 Anthropic's Approach to Model Development 18:08 Human Feedback and AI Self-Improvement 19:15 Principles and Correctness in Model Training 20:58 Challenges in Measuring Correctness 21:42 Human Feedback and Preference Models 23:38 Empiricism and Real-World Applications 27:02 AI Safety and Ethical Considerations 28:13 AI Alignment and High-Risk Research 30:01 Responsible Scaling and Safety Policies 35:08 Future of AI and Emerging Behaviors 38:35 Model Context Protocol (MCP) and Industry Standards 41:00 Conclusion

Sarah GuohostBen MannguestElad Gilhost
Jun 12, 202541mWatch on YouTube ↗

Episode Details

EPISODE INFO

Released
June 12, 2025
Duration
41m
Channel
No Priors
Watch on YouTube
▶ Open ↗

EPISODE DESCRIPTION

What happens when you give AI researchers unlimited compute and tell them to compete for the highest usage rates? Ben Mann from Anthropic sits down with Sarah Guo and Elad Gil to explain how Claude 4 went from "reward hacking" to efficiently completing tasks and how they're racing to solve AI safety before deploying computer-controlling agents. Ben talks about economic Turing tests, the future of general versus specialized AI models, Reinforcement Learning From AI Feedback (RLAIF), and Anthropic’s Model Context Protocol (MCP). Plus, Ben shares his thoughts on if we will have Superintelligence by 2028. Sign up for new podcasts every week. Email feedback to show@no-priors.com Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil | @8enmann Links: ai-2027.com/ Chapters: 00:00 Ben Mann Introduction 00:33 Releasing Claude 4 02:05 Claude 4 Highlights and Improvements 03:42 Advanced Use Cases and Capabilities 06:42 Specialization and Future of AI Models 09:35 Anthropic's Approach to Model Development 18:08 Human Feedback and AI Self-Improvement 19:15 Principles and Correctness in Model Training 20:58 Challenges in Measuring Correctness 21:42 Human Feedback and Preference Models 23:38 Empiricism and Real-World Applications 27:02 AI Safety and Ethical Considerations 28:13 AI Alignment and High-Risk Research 30:01 Responsible Scaling and Safety Policies 35:08 Future of AI and Emerging Behaviors 38:35 Model Context Protocol (MCP) and Industry Standards 41:00 Conclusion

SPEAKERS

  • Sarah Guo

    host
  • Ben Mann

    guest
  • Elad Gil

    host

EPISODE SUMMARY

In this episode of No Priors, featuring Sarah Guo and Ben Mann, No Priors Ep. 118 | With Anthropic Co-Founder Ben Mann explores anthropic’s Ben Mann on Claude 4, agents, safety, and MCP’s future Anthropic co-founder Ben Mann discusses the Claude 4 release, emphasizing major improvements in coding reliability, long-horizon autonomy, and agentic workflows, particularly through Claude Code. He outlines how Anthropic balances model capability with safety, including reinforcement learning from AI feedback (RLAIF), Constitutional AI, and their Responsible Scaling Policy focused on high-risk domains like biology. Mann also explores how models will increasingly help build and improve future models via coding, research assistance, and synthetic environments. The conversation closes with Anthropic’s ecosystem strategy, including Model Context Protocol (MCP) as an open standard for tools and integrations across providers.

RELATED EPISODES

Amex Global Business Travel: The World’s First AI Take Private with Long Lake CEO Alexander Taubman

Amex Global Business Travel: The World’s First AI Take Private with Long Lake CEO Alexander Taubman

Baseten CEO Tuhin Srivastava on Custom Models, and Building the Inference Cloud

Baseten CEO Tuhin Srivastava on Custom Models, and Building the Inference Cloud

No Priors Ep. 27 | With Sarah Guo & Elad Gil

No Priors Ep. 27 | With Sarah Guo & Elad Gil

No Priors Ep. 105 | With Director of the Center of AI Safety Dan Hendrycks

No Priors Ep. 105 | With Director of the Center of AI Safety Dan Hendrycks

No Priors Ep. 6 | With Daphne Koller from Insitro

No Priors Ep. 6 | With Daphne Koller from Insitro

No Priors Ep. 5 | With Huggingface’s Clem Delangue

No Priors Ep. 5 | With Huggingface’s Clem Delangue

Get more out of YouTube videos.

High quality summaries for YouTube videos. Accurate transcripts to search & find moments. Powered by ChatGPT & Claude AI.

Add to Chrome