
Mark Zuckerberg — AI will write most Meta code in 18 months
Mark Zuckerberg (guest), Dwarkesh Patel (host)
In this episode of Dwarkesh Podcast, featuring Mark Zuckerberg and Dwarkesh Patel, Mark Zuckerberg — AI will write most Meta code in 18 months explores zuckerberg predicts AI-coded Meta, open-source dominance, and weirder internet Mark Zuckerberg discusses Meta’s Llama 4 roadmap, emphasizing open-source models, multimodality, and an upcoming two-trillion-parameter ‘Behemoth’ model designed mainly as a teacher for smaller distilled systems.
Zuckerberg predicts AI-coded Meta, open-source dominance, and weirder internet
Mark Zuckerberg discusses Meta’s Llama 4 roadmap, emphasizing open-source models, multimodality, and an upcoming two-trillion-parameter ‘Behemoth’ model designed mainly as a teacher for smaller distilled systems.
He predicts that within 12–18 months, most of the code for Meta’s AI efforts will be written by AI agents, while arguing this will increase, not decrease, overall demand for human work and new services.
Zuckerberg frames Meta’s north star as a ubiquitous, personalized AI assistant embedded across messaging, apps, and AR glasses, and believes AI will reshape not just productivity and research but also entertainment, culture, and online social interaction.
He defends Meta’s semi-open Llama licensing, stresses the geopolitical stakes of AI standards and infrastructure, and reflects on content moderation, governance, and the need to avoid over-deference to government and media critics.
Key Takeaways
AI will soon write most frontier AI code, but within real-world constraints.
Zuckerberg expects that in 12–18 months, AI agents at Meta will generate the majority of code for Llama and related efforts—running tests, finding bugs, and exceeding average engineer quality—yet he stresses progress is bottlenecked by compute, infrastructure, and experimentation bandwidth, not just coding speed.
Get the full analysis with uListen AI
Open-source models are becoming the default, but Meta wants to steer the standard.
Zuckerberg argues 2024 is the year open-source broadly overtakes closed in usage, with multiple strong families (not just Llama), but he believes Meta’s consistent commitment is crucial; others may be ‘dabbling’ and could re-close when convenient, so Meta keeps pushing open models that encode ‘American’ values and secure architectures.
Get the full analysis with uListen AI
Benchmarks matter less than real user value and latency in products.
He notes Llama 4 Maverick can be pushed to the top of Chatbot Arena via tuning, but Meta chooses to optimize for Meta AI usage metrics, low latency, and cost per intelligence—especially for consumer and voice use cases—rather than gaming public leaderboards or chasing narrow reasoning benchmarks alone.
Get the full analysis with uListen AI
Distillation from giant models and even rival models will be central.
The two-trillion-parameter Behemoth is mainly a teacher: Meta intends to distill ~90–95% of its intelligence into 10x smaller, cheaper models, potentially combining strengths from multiple sources (e. ...
Get the full analysis with uListen AI
Personal AI will be pervasive, multimodal, and conversational throughout the day.
Zuckerberg envisions people constantly interacting with AI—within WhatsApp, Instagram, the standalone Meta AI app, and AR glasses—using full-duplex voice, persistent memory, and context from social graphs and feeds, making AI feel like an always-available, personalized collaborator and companion.
Get the full analysis with uListen AI
AI is likely to increase demand for human work by unlocking new services.
Using examples like customer support and self-driving, he argues that when AI cuts the cost of a service by an order of magnitude, entirely new offerings (e. ...
Get the full analysis with uListen AI
The future internet will be far more interactive, personalized, and ‘weird.’
He expects AI to supercharge culture: feeds evolving from static video to interactive, game-like AI experiences; ubiquitous co-created memes and media; and AI companions helping people who lack enough close friends—all making online life funnier, richer, and more idiosyncratic, while raising questions about reward hacking and healthy engagement.
Get the full analysis with uListen AI
Notable Quotes
“I would guess that, like, sometime in the next 12 to 18 months, we'll reach the point where, like, most of the code that's going towards these efforts is written by AI.”
— Mark Zuckerberg
“If you think that something someone is doing is bad, and they think it's really valuable, most of the time, in my experience, they're right and you're wrong.”
— Mark Zuckerberg
“I tend to think that, for at least the foreseeable future, this is gonna lead towards more demand for people doing work, not less.”
— Mark Zuckerberg
“I would guess that the world is gonna get a lot more... a lot funnier and, like, weirder.”
— Mark Zuckerberg
“These models encode values and ways of thinking about the world.”
— Mark Zuckerberg
Questions Answered in This Episode
If AI agents truly surpass most human programmers at Meta, how will the company redefine what its engineers do and how they’re trained or hired?
Mark Zuckerberg discusses Meta’s Llama 4 roadmap, emphasizing open-source models, multimodality, and an upcoming two-trillion-parameter ‘Behemoth’ model designed mainly as a teacher for smaller distilled systems.
Get the full analysis with uListen AI
What mechanisms could reliably measure ‘user value’ across different AI products so that labs don’t end up optimizing for metrics that hide subtle harms or biases?
He predicts that within 12–18 months, most of the code for Meta’s AI efforts will be written by AI agents, while arguing this will increase, not decrease, overall demand for human work and new services.
Get the full analysis with uListen AI
How should societies decide which cultural values and safety assumptions are embedded in foundational models that may become de facto global standards?
Zuckerberg frames Meta’s north star as a ubiquitous, personalized AI assistant embedded across messaging, apps, and AR glasses, and believes AI will reshape not just productivity and research but also entertainment, culture, and online social interaction.
Get the full analysis with uListen AI
Where is the line between AI companionship that fills a real social gap and AI that deepens isolation or ‘reward hacks’ people into preferring virtual relationships?
He defends Meta’s semi-open Llama licensing, stresses the geopolitical stakes of AI standards and infrastructure, and reflects on content moderation, governance, and the need to avoid over-deference to government and media critics.
Get the full analysis with uListen AI
In a world where enormous models like Behemoth are mainly teachers for distilled models, who should be responsible—and liable—for security and hidden backdoors in distilled systems that mix multiple upstream sources?
Get the full analysis with uListen AI
Transcript Preview
(instrumental music plays) I would guess that the world is gonna get a lot more... a lot funnier and, like, weirder. If you think that something someone is doing is bad, and they think it's really valuable, most of the time, in my experience, they're right and you're wrong.
I am worried that we're just removing all the friction between getting totally reward hacked by our technology.
We are trying to build a coding agent that advances LLaMA research. I would guess that, like, sometime in the next 12 to 18 months, we'll reach the point where, like, most of the code that's going towards these efforts is written by AI. I tend to think that, for at least the foreseeable future, this is gonna lead towards more demand for people doing work, not less. If you've gotten the cost of providing that service down to one-tenth of what it would have otherwise been, maybe now that actually makes sense to go do.
All right, Mark, thanks for coming on the podcast again.
Yeah, happy to do it. Good to see you.
You too. Last time you were here, you, um, you had launched LLaMA 3.
Yeah.
Now you've launched LLaMA 4.
Well, the first version.
That's right. What's new? What's exciting? What's changed?
Oh, well, I mean, uh, the whole field's so dynamic, so, I mean, I, I feel like a ton has changed since the last time that we talked. Um, Meta AI has almost a billion people using it now, monthly, so that's, um, that's pretty wild. Um, and, you know, I think that this is gonna be a really big year on all of this, because, um, especially once you start getting the personalization loop going, uh, which we're just starting to build in now really, um, from both the context that all the algorithms have about what you're interested in feed and all your profile information, all the social graph information, but also just what you're interacting with the AI about. I think that's just gonna be kind of the next thing that's, um, that's gonna be super exciting, so really big on that. The modeling stuff continues to make really impressive advances too, as, as you know. Um, the LLaMA 4 stuff, uh, I th- I'm pretty happy with the first set of releases. You know, we announced the, um, we announced four models, and we released the first two, the Scout and Maverick ones, which are kind of like the mid-size models, mid-size to small. Um, it's not like... You know, actually, the most popular LLaMA 3 model was, um, was the, the eight billion parameter model.
Mm.
So we're, we, we've, we've got one of those coming in the LLaMA 4 series too. Um, our internal code name for it is Little Llama.
(laughs)
But, um, but that, that's, that's, that's coming probably, you know, over, over the next, over the coming months. But the, um, the Scout and Maverick ones, um, you know, I mean, they're good. They're, they're some of the highest intelligence per cost that you can get of any model that's out there, natively multimodal, very efficient, run on one host, um, designed to just be very efficient and low latency for a lot of the use cases that we're building for internally, and, you know, that's our whole thing. We, we basically build what we're, what we want, and then we open source it so other people can use it too. So I'm excited about that. Um, I'm also excited about the Behemoth model, which is, is coming up. Um, that's gonna be our first model that is, uh, sort of at the frontier. I mean, it's like more than two trillion parameters, so it is... Y- y- I mean, it's, you know, as, as the name says, it's g- it's quite, quite big, um, so we're kind of trying to figure out how we make that useful for people. It's so big that we've had to build a bunch of infrastructure, um, just to be able to push train it ourselves, and we're kind of trying to wrap our head around how does the, like, like, the average developer out there, how are they gonna be able to use something like this, and how do we make it so it can be useful for, um, distilling into models that are of reasonable size to run? 'Cause you're, you're obviously not gonna wanna run, um, you know, something like that in a, in a consumer model. But, um, but yeah. I mean, it's, there, there's a lot to go. I mean, as, as you saw with the, with the LLaMA 3 stuff last year, the initial LLaMA 3 launch was, um, was exciting, and then we just kind of built on that over the year. 3.1 was when we released the 405 billion model. 3.2 is when we got all the multimodal stuff in. Um, so I, we basically have a roadmap like that for this year too, so a lot going on.
Install uListen to search the full transcript and get AI-powered insights
Get Full TranscriptGet more from every podcast
AI summaries, searchable transcripts, and fact-checking. Free forever.
Add to Chrome