Skip to content
No PriorsNo Priors

No Priors Ep. 40 | With Arthur Mensch, CEO Mistral AI

Open Source fuels the engine of innovation, according to Arthur Mensch, CEO and co-founder of Mistral AI. Mistral is a French AI company which recently made a splash with releasing Mistral 7B, the most powerful language model for its size to date, and outperforming much larger models. Sarah Guo and Elad Gil sit down with Arthur to discuss why open source could win the AI wars, their $100M+ seed financing, the true nature of scaling laws, why he started his company in France, and what Mistral is building next. Arthur Mensch is Chief Executive Officer and co-founder of Mistral AI. A graduate of École Polytechnique, Télécom Paris and holder of the Master Mathématiques Vision Apprentissage at Paris Saclay, he completed his thesis in machine learning for functional brain imaging at Inria (Parietal team). He spent two years as a post-doctoral fellow in the Applied Mathematics department at ENS Ulm, where he carried out work in mathematics for optimization and machine learning. In 2020, he joined DeepMind as a researcher, working on large language models, before leaving in 2023 to co-found Mistral AI with Guillaume Lample and Timothee Lacroix. 00:00 - Why he co-founded Mistral 04:22 - Chinchilla and Proportionality 06:16 - Mistral 7b 09:17 - Data and Annotations 10:33 - Open Source Ecosystem 17:36 - Proposed Compute and Scale Limits 19:58 - Threat of Bioweapons 23:08 - Guardrails and Safety 29:46 - Mistral Platform 31:31 - French and European AI Startups

Sarah GuohostArthur MenschguestElad Gilhost
Nov 8, 202332mWatch on YouTube ↗

At a glance

WHAT IT’S REALLY ABOUT

Mistral CEO Arthur Mensch Champions Efficient, Open-Source Frontier AI Models

  1. Arthur Mensch, CEO and co-founder of Mistral AI, explains how his team leverages a decade of optimization and scaling-law research to build highly efficient, small open-source language models like Mistral 7B. He argues that careful data curation, compression, and attention to inference cost can deliver models that run cheaply on commodity hardware while remaining surprisingly capable. Mensch strongly defends open source as essential for scientific progress and safety, criticizing current regulatory narratives around AI risk—especially bioweapons and arbitrary compute thresholds—as largely unsubstantiated and prone to regulatory capture. He outlines Mistral’s modular approach to safety and guardrails, its plans for larger models and agents, and why Europe, particularly France, is well-positioned to host a major global AI company.

IDEAS WORTH REMEMBERING

5 ideas

Optimize both training and inference to make AI economically usable at scale.

Mensch stresses that frontier models must be designed not only for raw benchmark performance but for low inference cost, enabling agents and ubiquitous deployment without prohibitive runtime expenses.

Small, well-trained models can be far more capable than expected.

By applying improved scaling laws and compression insights, Mistral 7B shows that a 7B-parameter model can be both fast and useful, running on devices like a MacBook Pro while matching or surpassing larger models on many tasks.

High-quality data curation is as critical as algorithmic innovation.

Mistral invests heavily in selecting and cleaning open web data for pre-training, treating data quality as a primary driver of model performance, distinct from later-stage instruction tuning.

Open sourcing current LLMs likely does not materially increase misuse risk.

Mensch argues there is no solid evidence that LLMs provide more dangerous capabilities than search engines for tasks like bioweapons, nor that knowledge access is the bottleneck for such misuse; thus blanket restrictions on open source are scientifically unfounded.

Safety should be implemented as modular guardrails, not baked-in censorship.

He advocates shipping raw models plus configurable filters for inputs and outputs (e.g., for hate speech, pornography), letting application builders and specialized safety providers compete to offer the best guardrailing solutions.

WORDS WORTH SAVING

5 quotes

We realized that there was also a lot of opportunity in actually compressing models more… with Mistral 7B we were definitely far away from the limit of compression.

Arthur Mensch

By doing what we do, by being much more open about the technology we create, we want to steer the community into a regime where things just work better, where things are safer because of more scrutiny.

Arthur Mensch

Nothing is showing that a LLM is actually marginally better than a search engine to find knowledge on topics that would enable bad use.

Arthur Mensch

Assuming that the model should be well behaved is, I think, a wrong assumption. You need to make the assumption that the model should know everything and then on top of that have some modules that moderate and guardrail the model.

Arthur Mensch

I’m not too worried about existential risk… There’s no evidence whatsoever that we are on the way of making that happen.

Arthur Mensch

Mensch’s research background in optimization, retrieval, mixture-of-experts, and scaling laws (including Chinchilla)Design and significance of small, efficient models like Mistral 7BData quality, pre-training vs. instruction tuning, and annotationsOpen-source AI as a scientific, economic, and safety choiceDebates around AI safety: content moderation, physical risk, and existential riskRegulation, compute thresholds, and the bioweapon narrativeMistral’s platform strategy, guardrails architecture, and European AI ecosystem

High quality AI-generated summary created from speaker-labeled transcript.

Get more out of YouTube videos.

High quality summaries for YouTube videos. Accurate transcripts to search & find moments. Powered by ChatGPT & Claude AI.

Add to Chrome