Skip to content
Lenny's PodcastLenny's Podcast

Mike Krieger: How Claude writes 90% of Anthropic's code

At Anthropic, Krieger had to rearchitect the merge queue itself: with Claude writing 90% of new code and 70% of pull requests, the review pipeline broke.

Lenny RachitskyhostMike Kriegerguest
Jun 5, 20251h 6mWatch on YouTube ↗

EVERY SPOKEN WORD

  1. 0:004:25

    Introduction to Mike Krieger

    1. LR

      ... 90% of your code roughly is written by AI now.

    2. MK

      The team that works in the most futuristic way is the Claude code team. They're using Claude code to build Claude code in a very self-improving kind of way. We really rapidly became bottlenecked on other things, like our merge queue. We had to completely re-architect it because so much more code was being written and so many more pull requests were being submitted. Over half of our pull requests are Claude Code generated. Probably at this point it's probably over 70%. And it just completely blew out the expectations of it.

    3. LR

      You guys are at the edge of where things are heading.

    4. MK

      I had the very bizarre experience of I had two tabs open. It was AI 2027 and my product strategy, and it was this, like, moment where I'm like, "Wait, am I the character in the story?"

    5. LR

      It feels like ChatGPT is just winning in consumer mindshare. How does that inform the way you think about product, strategy, and mission?

    6. MK

      I think there's room for several generationally important companies to be built in AI right now. How do we figure out what we wanna be when we grow up versus, like, what we currently aren't or wish that we were, or, like, see other players in the space being?

    7. LR

      What's something that you've changed your mind about what AI is capable of and where AI is heading?

    8. MK

      I had this notion coming in, like, yes, these models are great, but are they able to have an independent opinion? And it's actually really flipped for me only in the last month.

    9. LR

      Today my guest is Mike Krieger. Mike is chief product officer at Anthropic, the company behind Claude. He's also the co-founder of Instagram. He's one of my most favorite product builders and thinkers. He's also now leading product at one of the most important companies in the world, and I'm so thrilled to have had a chance to chat with him on the podcast. We chat about what he's changed his mind about most in terms of AI capabilities in the years since he joined Anthropic, how product development changes and where bottlenecks emerge when 90% of your code is written by AI, which is now true at Anthropic. Also his thoughts on OpenAI versus Anthropic, the future of MCP, why he shut down Artifact, his last startup, and how he feels about it, also what skills he's encouraging his kids to develop with the rise of AI, and we close the podcast on a very heartwarming message that Claude wanted me to share with Mike. A big thank you to my newsletter Slack community for suggesting topics for this conversation. If you enjoy this podcast, don't forget to subscribe and follow it in your favorite podcasting app or YouTube. Also, if you become an annual subscriber of my newsletter, you get a year free of a bunch of incredible products, including Linear, Superhuman, Notion, Perplexity, and Granola. Check it out at lennysnewsletter.com and click bundle. With that, I bring you Mike Krieger. This episode is brought to you by Productboard, the leading product management platform for the enterprise. For over 10 years, Productboard has helped customer-centric organizations like Zoom, Salesforce, and Autodesk build the right products faster. And as an end-to-end platform, Productboard seamlessly supports all stages of the product development life cycle, from gathering customer insights, to planning a roadmap, to aligning stakeholders, to earning customer buy-in, all with a single source of truth. And now, product leaders can get even more visibility into customer needs with Productboard Pulse, a new voice of customer solution. Built-in intelligence helps you analyze trends across all of your feedback and then dive deeper by asking AI your follow-up questions. See how Productboard can help your team deliver higher impact products that solve real customer needs and advance your business goals. For a special offer and free 15-day trial, visit productboard.com/lenny. That's productboard.com/L-E-N-N-Y. Last year, 1.3% of the global GDP flowed through Stripe. That's over $1.4 trillion. And driving that huge number are the millions of businesses growing more rapidly with Stripe. For industry leaders like Forbes, Atlassian, OpenAI, and Toyota, Stripe isn't just financial software. It's a powerful partner that simplifies how they move money, making it as seamless and borderless as the internet itself. For example, Hertz boosted its online payment authorization rates by 4% after migrating to Stripe. And imagine seeing a 23% lift in revenue like Forbes did just six months after switching to Stripe for subscription management. Stripe has been leveraging AI for the last decade to make its product better at growing revenue for all businesses, from smarter checkouts, to fraud prevention, and beyond. Join the ranks of over half of the Fortune 100 companies that trust Stripe to drive change. Learn more at stripe.com.

  2. 4:257:43

    What Mike has changed his mind about regarding AI capabilities

    1. LR

      Mike, thank you so much for being here and welcome to the podcast.

    2. MK

      I'm really happy to be here. I've been looking forward to this for a while.

    3. LR

      Wow. I- I love to hear that. I've also been looking forward to this for a while. Uh, I have so much to talk about. So first of all, you've been at Anthropic for just over a year at this point. Congrats, by the way, on hitting, hitting the cliff.

    4. MK

      Thank you. (laughs)

    5. LR

      (laughs)

    6. MK

      Not that we're tracking.

    7. LR

      That's right. So let me just ask you this. So you've been at Anthropic for about a year. What's something that you've changed your mind about from before you joined Anthropic to today about what AI is capable of and where AI is heading?

    8. MK

      Two things. One is, like, a pace and timeline question, and the other one is a capability question. So maybe I'll take the second one first. I had this notion coming in, like, yes, these models are great. They're gonna be able to produce code. They're gonna be able to, you know, write, you know, hopefully in your voice eventually. But are they able to sort of have an independent opinion? And it's actually really flipped for me only in the last month and only with Opus 4, where my go-to product strategy partner is Claude, and it has been basically for that full year, where I'll write an initial strategy. I'll share it with Claude, basically, and I'll have it, you know, look at it. And in the past, it's pretty anodyne kind of comments that it would leave, like, "Oh, have you thought about this?" And it's like, "Yeah, yeah, I thought about that." And Opus 4, I was working on some strategy for our second half of the year, was the first one. It was, like, Opus 4 combined with our advanced research, but it really went out for a while and it came back and I was like, "Damn, you really looked at it in a new way." And so that's, like, a thing that I've maybe... I didn't feel like it would never be able to do that, but I wasn't sure how soon it'd be able to, like, come up with something where I look at it and I'm like, "Yep, that, that is a new angle that I hadn't been looking at before and I'm going to incorporate that immediately into how, how I think about it." So that's probably the, the biggest shift that I've had is, like, inde-I don't know about independence is the right word, but like creativity and sort of novelty of thought relative to how I'm, I'm thinking about things. And then the timeline one, it's like so interesting because, you know, uh, I was sitting next to Dario yesterday and he's like, "I keep making these predictions and people keep laughing at me and then they come true." And it's like... And it's funny to have this happen over and over again. And he's like, "Not all of them are gonna be right," you know. But even, I think as of last year, he was talking about, you know, we're up 50% on SWE Bench, which is this like, you know, benchmark around how well the models are at, at coding. Uh, he's like, "I think we'll be at 90% by the end of 2025," or something like that. And sure enough, we're at about 72 now with the new models, and we were at 50% when he made that prediction. And it's like, continued to scale pretty much like as predicted. And so I've taken the timelines a lot more seriously now, and... And if you read AI 2027 that like, you know-

    9. LR

      I, I have.

    10. MK

      Yeah.

    11. LR

      It was (laughs) it was made by HeartRace.

    12. MK

      Yeah, and I had the very bizarre experience of I had two tabs open, it was AI 2027 and my product strategy, and it was this like moment where I'm like, "Wait, am I the character in the story?" Like is this... Or how much is this converging? But, you know, you read that and you're like, "Oh, 2027, that's like, that's years away." But you're like, no, mid-2025, and like things continue to, uh, to improve and the models continue to be able to do more and more, and they're able to act agenitically and they're able to have memory and they're able to act over time. So I think my, like, my confidence in the timelines, and I don't know exactly how they manifest, have definitely just solidified over the last year.

  3. 7:439:00

    How to avoid scary AI scenarios

    1. MK

    2. LR

      Wow. (clears throat) Uh, I, I wasn't expecting to go down that, 'cause that, that, that paper was scary. And I'm curious just, I guess, l- I can't help but ask, just thoughts on just how do we avoid the scary scenario that that paper paints of where AI getting really smart goes?

    3. MK

      Yeah. I mean, I, uh, I... This maybe ties into like, I've been here a year, like why did I join Anthropic? I was watching the models get better and even, you know, you could see it in, in '24 and like, you know, early 2024. And looking at my kids, I'm like, all right, they're gonna grow up in a world with AI. It's un- unavoidable. What is the thing that I can... Like where can I maximally apply my time to, like nudge things towards going well? And I mean, that's a lot of what people think about across the industry, especially at Anthropic. And so I think, you know, coming to an agreement and a shared framework and understanding of like, what does going well look like? What is the kind of human-AI relationship that we want? How will we know along the way? What do we need to build and develop and research along the way? I think those are all the kinda key questions. And, you know, some of those are product questions, and, and some of those are, are research and interpretability questions. But for me it was like the, the strongest reason to join was, okay, I think there's a, there's a lot of contribution that Anthropic can have around, like nudging things to go better, and if I can have a part to play there, like let's do it.

  4. 9:0011:58

    Skills kids will need in an AI world

    1. MK

    2. LR

      I, I love that answer. Uh, speaking of kids, so you've got two kids. I've got a young kid, he's, uh, just about to turn two. I'm curious just what skills you're encouraging your kids to build as this, you know, AI becomes more and more of our future and some jobs, you know, will be changed, and just what do you, what do you... What advice do you have?

    3. MK

      We have this, uh, you know, breakfast, we eat breakfast with the kids every morning, and sometimes some question will come up, you know, like, you know, something about, like physics. And our oldest kid's almost six, but, you know, they, they ask like funny questions about like, you know, uh, you know, the solar system or physics or, you know, in a six-year-old way. And before we reach for Claude, 'cause at first, you know, my instinct is like, "Oh, I wonder how Claude will do at this question." And like we started changing it to like, "Well, how would we find out?" You know? And the answer can't just be, "We'll ask Claude," you know? So all right, like, well, we could do this experiment, we could have this thing. So I think nurturing curiosity and like still having a sense of... I don't know, the scientific process sounds grandiose to instill in like a six-year-old, but like, that process of like discovery and asking questions and then, you know, systematically working your way through it I think will still be important. And of course AI will be an incredible tool for helping like resolve large parts of that, but that process of inquiry I think is still really important and independent thought. My favorite moment with my kid, uh, 'cause they're... she's very headstrong, our six-year-old, she's... You know, I was like... She said something and I was like, "I wasn't sure if it was true." It was, um, uh... Oh, it was like coral as a, as an animal or like coral is alive, I don't even remember what the details of it, and I was like, "I don't know if that's true." And she's like, "It's definitely true, Dad." I'm like, "All right," like, "Let's ask Claude on this one." And she's like, "You can ask Claude, but I know I'm right," and I'm like, "I love that." Like, I want that kind of level of, you know, not just sort of, uh, delegating all of your cognition to the, you know, to the AI, 'cause it won't always get it right. And also, uh, kind of like s- you know, kind of short circuits any kind of independent thought. So the skill of, of asking questions, inquiry, uh, and independent thinking, I think those are all the pieces. What that looks like from a, like job or occupation perspective, like I'm just keeping an open mind and I'm sure that'll radically change between, between now and then.

    4. LR

      It's interesting. I had Tobi Lutke, uh, Shopify CEO on the podcast, and he had the same answer for what he's encouraging his kids to, uh, to develop is curiosity. And, uh, and so it's interesting that's a common thread.

    5. MK

      The, you know, K through eight school our kid goes to had a, an AI sort of AI and education expert come in, and I had a very low bar or like very low expectations of what this conversation was gonna be like. And actually, I think it went over most of the people, uh, in the heads... the audience's heads 'cause he was like, "All right, well let me take you all the way back to Claude Shannon and information theory," and I could see people's eyes going like, "What did I like sign up for and why am I here in this like school auditorium hearing about, you know, information theory?" But he did a really nice job, I think, of also just imagining like, you know, there will be different jobs and we don't know what those jobs are going to be, and so like what are the skills and techniques and, and, and remain open-mindedness and around like what the... what, what the exact way we recombine those things, and even those will probably change three times between now and eight- when they're

  5. 11:5817:12

    How product development changes when 90% of code is written by AI

    1. MK

      18.

    2. LR

      I want to go back to... So we're talking about timelines and how things are changing. So I've seen these stats that you've shared, other folks at Anthropic have shared, about how much of your code is now written by AI. So people have shared stats from like 70% to like 90%. There was an engineer lead that shared like 90% of your code roughly is written by AI now, which first of all is just insane.... that, like, it went from zero to 90%, I don't know, a few years? Something like that?

    3. MK

      Yeah. Basically.

    4. LR

      I don't think that's- I don't think people are talking about this enough. That's just wild. You guys are basically at the bleeding edge. I've never heard of a company that has this high a percentage of code being written by AI. So, you guys are at the edge of where things are heading. I think most companies will get here. How has product development changed knowing so much of your code is now written by AI? So, usually it's like PM is like, "Here's what we're building," engineer builds it, ships it. Is it still kind of roughly that or is it now PMs are just going straight to Claude, "Build this thing for me," engineers are doing different things? Just what looks different in a world where 90% of your code is written by AI?

    5. MK

      Yeah. It- it's really interesting 'cause I think the, like, the role, like, the role of engineering has changed a lot. But the, the kind of suite of people that come together to produce a product hasn't yet. And I think for the worse in a lot of ways because I- I think we're still holding on some assumptions. So, I think they're- the, the roles are still fairly similar, although we'll now get, and my favorite things that happen now are sometimes PMs that have an idea that they want to express, or designers that have an idea they want to express, will use Claude and, like, maybe even artifacts to, like, put together an actual, like, functional demo, and that has been very, very helpful. Like, no, no, this is what I mean, like, and that- that makes it tangible. That's probably the biggest, like, role shift is, like, prototyping happening earlier in the process be a more of this kind of, you know, uh, you know, code plus design piece. What I've learned though is, like, the process of knowing what to ask the AI, how to compose the question, how to even think about, like, structuring a change between the backend and the frontend, those are still very difficult and specialized skills, and they still require the engineer to think about it. And we really rapidly became bottlenecked on other things, like our merge queue, which is the sort of, sort of get in line to get your change accepted by, uh, you know, the, the, the system that then deploys it to production. We had to completely rearchitect it because so much more code was being written and so many more pull requests were being submitted that it just completely blew out the expectations of it. And so it's like, I don't know if you've ever read, is it The Goal? The classic, like, process optimization book.

    6. LR

      Mm-hmm. Yeah.

    7. MK

      And you realize there's like this, like, critical path theory. I've just found all these new bottlenecks in our system, you know. There's an upstream bottleneck which is decision-making and alignment. And a lot of things that I'm thinking about right now is like, how do I provide the, like, minimum viable strategy to let people feel empowered to go run and prototype and build and explore at the edge of model capabilities? I don't think I've gotten that right yet, but it's something I'm working on. And then once the, uh, building is happening, other bottlenecks emerge like, let's make sure we don't step on each other's toes. Let's think through all the edge cases here ahead of time so that we're not blocked on the engineering side. And then when the work is complete and we're getting ready to, to, to ship it, what are all those bottlenecks as well? Like, let's do the air traffic control of landing the change, like, how do we figure out launch strategy? So, I think we're- the- there hasn't been as much pressure on changing those until this year, but I, I would expect that, like, a year from now, the way that we are, like, conceiving of building and shipping software just changes a lot because it's gonna be very painful to do it the current way.

    8. LR

      Wow. That is extremely interesting. So, it used to be here's an idea, let's go design it, build it, ship it, mer- merge it, and then ship it. And usually the bottleneck was engineering taking time to build the thing and then design. And now you're saying the two bottlenecks you're finding are, okay, deciding what to build and aligning everyone. And then it's actually, like, the queue to merge it into production and, uh, and, and I imagine review it too is probably a part of it.

    9. MK

      Reviewing has really changed too, and in- in- in many ways our most, uh, perhaps unsurprisingly, the team that works in the most futuristic way is the Claude Code team 'cause they're using Claude Code to build Claude Code in a very self-improving kind of way. And, you know, early on in that project, they would do very line by line pull request reviews, you know, in the way that you would for any other, you know, project. And they've just realized, like, Claude is generally right and it's producing, you know, pull requests that are probably larger than most people are gonna be able to review, so can you use a different Claude to review it and then do the human almost, like, acceptance testing more than trying to, like, review line by line? There's definitely pros and cons, and, like, so far it's gone well, but I could also imagine it going off the rails and then having, like, a completely both unmaintainable or even understandable by Claude code base that hasn't happened. But watching them, like, change their review processes definitely has, uh, has been, has been interesting. And yeah, like, the merge queue is one instance of the, of the kind of bottom bottleneck that forms down there, but there's other ones which is, how do we make sure that we're still, like, building something coherent and, like, packaging it up into, like, a moment that we can share with people? And whether that's around a launch moment, whether that's about, like, then enabling people to use this thing and, like, talking about it. Like, the, the classic things of building something useful for people and then making it known that you've built it and then learning from their feedback, like, still exists. We've just, like, made a portion of that whole process much more efficient.

    10. LR

      I heard you describe this as you guys are patient zero for this way of working.

    11. MK

      Yes.

  6. 17:1221:21

    Claude helping with product strategy

    1. MK

    2. LR

      I love that. Do you have a sense of what percentage of Claude Code is written by Claude Code?

    3. MK

      At this point, I would be shocked if it wasn't 95% plus. I'd have to ask Boris and the other tech leads on there. But what's been cool is, um, uh, so nitty-gritty stuff. Claude Code is written in TypeScript. It's actually our largest TypeScript project. Most of the rest of Anthropic is written in Python, some Go, um, some Rust now. But it's not, you know, we're not, like, a TypeScript shop. And so, uh, I saw a great comment yesterday in our Slack where somebody had this thing that was driving them crazy about Claude Code, and they're like, "Well, I don't know any TypeScript. I'm just gonna, like, talk to Claude about it and do it." And they went from that to pull request in an hour and solved their problem, and they, like, you know, submitted a pull request. And that kind of breaking down the barriers, one, it changes your sort of, um, uh, barrier to entry for any kind of, uh, kind of newcomer to the project. I think it can let you choose the right language for the right job, for example. I think that helps as well. But I think it, like, also just reinforces, like, Claude Code being that patient alpha of that, you know, where, like, e- contributions from outside the team can be Claude Coded as well.

    4. LR

      Wow.This is just, (clears throat) it's just gonna- can- c- continue to blow my mind, like all, (laughs) all these things that you're sharing. '95% of Claude code is written by Claude code, roughly.

    5. MK

      That's my guess.

    6. LR

      Uh...

    7. MK

      Yeah, I'd- I'll- I'll come back with the real stat, but it's, it, I mean, if you ask-

    8. LR

      Yeah.

    9. MK

      ... the team, that's how that they are working, and that's how they're getting contributions from across the company too.

    10. LR

      It's interesting, going back to your point about strategy being assisted by Claude itself, and your point about how a lot of the bottlenecks now are kind of the top of the funnel of coming up with ideas aligning everyone. It's interesting that Claude is already helping with that also, of helping you decide what to build. So if, if those two bottlenecks are aligning, deciding what to build and then just, like, merging and getting everything, where do you see the most, uh, interesting stuff happening to help you speed those things up?

    11. MK

      Yeah, I think that on that- on that first front, like, I started the year, um, by writing a doc that was effectively like, "What... How do we do product today, and where is Claude not showing up yet that it should?" And I think that upstream part is the next one that goes interesting. Like, at your conference, I talked to somebody who's working on, like, a PRD GPT, kind of like ChatPRD, I think was-

    12. LR

      ChatPRD.

    13. MK

      ... was, was the- yeah.

    14. LR

      Clear vote. Mm-hmm.

    15. MK

      Um, so, you know, can we push more on, you know, can Claude be a partner in figuring out what to build, what the market size is if you want to approach it that way, what the user needs are if you, if you look at a different way. Like, we think a lot about the virtual collaborator at Anthropic, and one of the ways in which I think that can show up is, "Hey, I'm in the Discord, the, you know, the- the Claude Anthropic Discord. I'm in the user fora. I'm on X and I'm reading things and, like, here's what's emergent." That's step one. Models can- can do that today. Step two, which the models probably can do today, we just have to wire them up to do it, is like, "And not only are there problems, here's, like, how I think you might be able to solve them," and then taking that through to, like, an AI, like, put together a pull request to, like, solve this thing that I'm seeing, like, feels very achievable this year, um, them stringing those things together. And we're limited more... This is why MCP is exciting to me. Like, we're limited more around, like, making sure the context flows through all of that, so we have the right access to those things more than the model's capability to- to reason and propose. Now, the model might not have, like, perfect UI taste yet, so there's definitely room for design to intervene and be like, "Oh, that's not quite how I would solve the problem of- of this not showing up," but I, you know, I would get very excited. I will give you a really, uh, small example, but we changed the, on Claude AI, uh, used to be able to just copy, uh, markdown from artifacts or code from artifacts, and we changed it so you can actually download it and- and export it. So we changed the button to export, and we got a bunch of feedback, like, "Well, how do I copy now?" And the answer is, like, you drop it down and it's copied. It's just, like, mind, you know, one of those things where it's, like, made sense, but we probably got it, like, not quite right. That feedback was in the- our UX channel. Like, I would have loved, like, an hour later for Claude to be like, "Hey, if we do want to change it back, here's the PR to do it. And by the way, eventually, and then I'm gonna spin up an AB test to see if this changes metrics, and then we'll see how it looks in a week." Like, this stuff feels... If you'd told me that about a year and a half ago, I'd be like, "Ah, yeah, maybe, like, '27, maybe, like, '26," but it's pretty m- like, I- it really feels, you know, just at the tip of capabilities right now.

    16. LR

      Wow.

  7. 21:2124:00

    A new way of working

    1. LR

      Okay, so (clears throat) you mentioned the Lenny and Friends Summit. I wanted to talk about this a bit. So you were on a panel with Kevin Weil, the CPO of OpenAI. I think it was the first time you guys did this. Maybe the last time for now.

    2. MK

      Yeah, we haven't done it since, not for any reason. I had a lot of fun.

    3. LR

      What a, what a legendary panel we assembled there with Sara Guo, uh, moderating. And you made this comment that actually ended up being the most rewatched part of the- of the interview, which is that you've kind of... You were putting product people on the model team and working with researchers making the model better, and you were putting some product people on the product experience, making the UX more intuitive, making all that better. And you found that almost all the leverage came from the product team working with the researchers.

    4. MK

      Yes.

    5. LR

      And so you've been doing more of that. So first of all, does that continue to be true? And second of all, what are the implications of that for product teams?

    6. MK

      It's continued to be true, and in- in fact, I think that the, if the proportion was already, like, skewing towards having more of that embedding, I've just become more and more convinced. Like, I have this... I- I didn't feel as strongly about it during your, you know, the summit, and now I feel really strongly about it, which is, if any- If we're shipping things that could have been built by anybody just using our models off the shelf... There's great stuff to be built by using our models off the shelf, by the way, don't get me wrong, but, like, where we should play and, like, what we can do uniquely should be stuff that's really at that, like, magic intersection between the two, right? Artifacts being a great example, and, uh, if you play with artifacts with- with Claude 4.0, that's an actually really interesting example where we took somebody from our, we have, we call it Claude Skills, which is a team that really is like doing the post-training around teaching Claude, you know, some of these, like, really specific skills, and we paired it with some product people. And then together, we revamped how this looks in the product today, and, like, what Claude can do way better than just like, "Yeah, we just, like, used the model, and we, like, prompted a little bit." Like, that's just not enough. We need to be in that, like, fine-tuning process. So, so much of what, you know, if you look at what we're working on right now, what we've shipped recently between, like, research and all these other things, like, are things that we... Like, the- the functional unit of work at Anthropic is no longer, like, take the model and then, like, go, like, work with design and product to go ship a product. It's more like we are at, like, we're in the post-training conversations around how these things should work, and then we are in the building process, and we're, like, feeding those things back and looping them back. Like, I think it's exciting. It's also, um, a new way of working that, like, not all PMs have, but the PMs that have the most sort of internal positive feedback from both research and engineering are the ones that get it, that, like-

    7. LR

      Mm-hmm.

    8. MK

      Uh, I was in a product review yesterday. I was like, "Oh, you know, if we wanna do this memory feature, like, we should talk to the eng- the researchers, because we just shipped a bunch of, like, memory capabilities in Claude 4.0." They're like, "Yeah, yeah, we've been talking to them for weeks. Like, this is how we're manifesting it." It's

  8. 24:0027:23

    The future value of product teams in an AI world

    1. MK

      like, "Okay, whew, I feel- feel good. I feel like we're doing the right things now."

    2. LR

      So let me pull on this thread, uh, more. There's something I've been thinking about along these lines. So essentially, there's, like, a big part of Anthropic that's building this super intelligent giga-brain that's gonna do all these things for us over time, and then there's, as you said, there's, like, the product team that's building the UX around the super intelligent giga-brain. And over time, this super intelligence is gonna be able to build its own stuff.And so, I guess just where do you think the most value will con- will come from pro- traditional product teams over time? I know that's different 'cause you guys are a foundational LM company and not most companies don't work this way, but just, I don't know, thoughts on just the where most value will come from product teams over time working on AI.

    3. MK

      I think there's still value, a lot of value in two things. One is making this all comprehensible. I think we've done an okay job. I think we could do a much better job of making this comprehensible. It's still, like, the difference between somebody who's really adept at using these tools in their work and most people is huge. And I mean, maybe that's the most literal answer to your earlier question around, like, what, what skills to learn. That is a skill to, to learn and use it in the same way that I remember, I, I redid like computer lab class when I was in like middle school. I remember being, like, really good at Google. And that was actually a skill back in the day, you know, like to think in terms of like, this information is out there. How do I query for it? How do I do it? And I think it actually was like a, a, an advantage at the time. Of course, now Google is pretty good at figuring out what you're trying to do if you like are only in the neighborhood and like there's less of that research kind of need. But I still think that's a necessary part of like good product development, which is like the capabilities are there. And even if the like, even if Claude can create products from scratch, what are you building and how do you make it comprehensible? Like, still hard because I think that like gets at like this much deeper empathy and like understanding of human needs and psychology. Like, I was a human computer interaction major. I've still been talking in my book here. Like, I still feel like that is a, a, a very, very, very, very necessary skill. So that's one. Two is, and this, you know, straight to call back to another one of your guests, like strategy, like how we win, where we'll play, like figuring out where exactly you're gonna want to like... Of all the things that you could be spending your time or your, uh, your tokens or your computation on, like what, what, what do you wanna actually go and do? You could be wider probably than you could before, but you can't do everything. And even like from an external perspective, if you're seen to be doing everything, like it's way less clear around like how you're, how you position yourself. So like strategy I think is still that, the second piece. And then the third one is opening people's eyes to what's possible, which is a continuation of making it understandable. But we were in a demo with a, a financial services company recently, and we were like working on like, here's how you can use our analysis tool and MCP together. And, and like you could see their eyes light up and you're like, "Ah, okay." Like there's still, so we call it overhang, right? Like the delta between what the models and the products can do and how it's been- they're being used on a daily basis, huge overhang. So that's where still like a, a very, very strong necessary role for product.

    4. LR

      Okay. That's an awesome answer. So essentially, areas for product teams to lean into more is strategy. Just getting better and better at strategy, figuring out what to build and how to win in the market, making it easier to help people understand how to leverage the power of these tools, so comprehensibility, and kind of along those lines is opening people's eyes to the potential of these sorts of things. That's where product can still help.

    5. MK

      Exactly.

    6. LR

      Awesome.

  9. 27:2329:57

    Prompting tricks to get more out of Claude

    1. LR

      So kind of along those lines actually, do you have any just like prompting tricks for people, things you've learned to get more out of Claude when you chat with it?

    2. MK

      Sometimes it, you know, it's funny because we, uh, in, in some ways, we have like the ultimate prompting job, which is to write the system prompt for Claudia. And we publish all of these, which I think is, is like a, you know, another nice area of transparency. And we are always careful when giving prompting advice because at least officially, but I'm, I'm gonna... I'll give you the unofficial version because like you don't want things to become like, uh, like we think this works, but we're not sure why, you know. But I, um, I'll do small things like in cloud code and we actually do react to this very literally. But I always like to ask it to like if I wanted to use more reasoning, like think hard and it'll like, you know, use it, use it kind of a different, uh, flow. And I always just start with that, you know. Um, nudging, there's a great essay around like make the other mistake. Like if you tend to be too nice, can you focus on like even if you're trying to be more critical or more blunt, you're probably not gonna be the most critical, blunt person in the world. Um, and so with Claude, sometimes I'm like, "Be brutal, Claude. Like roast me. Like tell me what's wrong with this strategy." I think, I know we were talking earlier about the, you know, Claude as thought partner around like critiquing product strategy. Uh, I think I previously would say things like, "You know, like what, what could be better on this product strategy?" And I'm just like, "You know, just roast this product strategy." And Claude's like a pretty nice, you know, entity. It's not gonna be... It, it's hard to push it to be super brutal, but it forces it to be a little bit more, uh, critical as well. The last thing I'll say is, so we have a team called Applied AI that does a lot of like work with our customers around optimizing Claude for their use case. And we basically took their insights and their way of working and we put it into a product itself. So if you go to our console, our workbench, we have this thing called the prompt improver where you describe the problem and you give it examples. And, uh, Claude itself will agentically create and then iterate on a prompt for you. I find what comes out of that ends up being quite different than what my intuitions would have been for a good prompt. And so I, I encourage folks to also check that out even for their own use cases, because while that tool is meant for an API developer putting a prompt into their product, it's equally applicable for, uh, a person doing a, a prompt for themselves. Like it'll insert XML tags, which no human is going to think to do ahead of time, but actually is very helpful for Claude to understand like what it should be thinking versus what it should be saying, et cetera. So that- that's another one is like watch our prompt improver and then note that like Claude itself is a very good prompter of Claude.

    3. LR

      Awesome. Okay. So we're gonna link to that, the prompt improver. The core piece of advice you shared earlier is just kind of do the opposite of what you would naturally do. So if you're like trying to be nice, just like be brutal. Be like very honest and frank with me. (laughs)

    4. MK

      Exactly. I find that worked quite well. Like what are the thought patterns that I've like fallen into that

  10. 29:5732:47

    The Rick Rubin collaboration on “vibe coding”

    1. MK

      you want to break me out of?

    2. LR

      Mm-hmm. I saw you guys just today maybe launched a Rick Rubin collab or something-

    3. MK

      Yes.

    4. LR

      ... in vibe coding. What's that all about? I don't-

    5. MK

      That was a, a, you know, what I heard about that... And then again, like this a lot of coalesced this week between model launch, developer event-

    6. LR

      Yeah. Big week.

    7. MK

      ... and the way of code. Um, we had our, our, one of our co-founders, Jack Clark is our, you know, head of policy and he got connected to Rick Rubin 'cause I think he's been thinking a lot about coding, the future of coding and creativity. And they've stayed in touch. And-You know, Rick got excited about this idea of, uh, like, he was creating, uh, like art and visualizations with Claude, and then he had these, like, ideas around, like, uh, the way of the vibe coder. And they put together this, actually, I love the, I mean, I love almost everything Rick Rubin, so like that, the aesthetic of it I think is just like, so on point too. But yeah, this sort of like, med- meditation's probably the right word, meditation on like creativity, working alongside AI, coupled with this like, uh, with this like really rich interesting visualizations. But it's just one of those things where like, uh, you know, internally we were like, oh yeah, and we're doing this like, Rick Rubin collab work. We're doing what? Like that is, that's amazing.

    8. LR

      I love the... I looked at it briefly and there's like that meme of him, like, just like thinking deeply sitting on a computer with a mouse-

    9. MK

      Yes, exactly.

    10. LR

      ... in like ASCII art, I think.

    11. MK

      It's totally, it's like ASCII art vibe.

    12. LR

      I'm excited to have Andrew Luo joining us today. Andrew is CEO of OneSchema, one of our longtime podcast sponsors. Welcome Andrew.

    13. NA

      Thanks for having me, Lenny. Great to be here.

    14. LR

      So what is new with OneSchema? I know that you work with some of my favorite companies like Ramp and Vanta and Watershed. I heard you guys launch a new data intake product that automates the hours of manual work that teams spend importing and mapping and integrating CSV and Excel files.

    15. NA

      Yes. So we just launched the 2.0 of OneSchema FileFeeds. We've rebuilt it from the ground up with AI. We saw so many customers coming to us with teams of data engineers that struggled with the manual work required to clean messy spreadsheets. FileFeeds 2.0 allows non-technical teams to automate the process of transforming CSV and Excel files with just a simple prompt. We support all of the trickiest file integrations, SFTP, S3, and even email.

    16. LR

      I can tell you that if my team had to build integrations like this, how nice would it be to take this off our roadmap and instead use something like OneSchema?

    17. NA

      Absolutely, Lenny. We've heard so many horror stories of outages from even just a single bad record in transactions, employee files, purchase orders, you name it. Debugging these issues is often like finding a needle in a haystack. OneSchema stops any bad data from entering your system and automatically validates your files, generating error reports with the exact issues in all bad files.

    18. LR

      I know that importing incorrect data can cause all kinds of pain for your customers and quickly lose their trust. Andrew, thank you so much for joining me. If you want to learn more, head on over to OneSchema.co. That's OneSchema.c-o.

  11. 32:4736:00

    How Mike was recruited to Anthropic

    1. LR

      Actually going back to kind of the beginning of your journey at Anthropic, what's the story of you getting recruited at Anthropic? Is there anything fun there?

    2. MK

      The, it all started, and I actually sent my friend this text. So Joel Lewenstein, who I've known, he actually, he and I built our first iPhone apps together in 2007 when the App Store was just out and you could still, you know, make money by selling dollar apps on the App Store, you know, back in the day. And we were, we were both at Stanford together and we were friends and we've stayed in touch over the years and we've never gotten to work together, uh, since then. We just like, we've just remained close. And you know, I was coming out of the Artifact experience. I was trying to figure out, do I start another company? I don't think so. I need a break from starting something from zero. Should I go work somewhere? I don't know, like what company would I want to go work at? And he reached out and he's like, "Look, I don't know if you'd at all consider joining something rather than starting something, but we're looking for a CPO. It would be, would you be interested in chatting?" And at that time, Claude 3 had just come out and I was like, "Okay, you know, like this company's clearly got a good research team. The product is so early still." And it was like, "Great, I'll take the, take the meeting." And I first met with Daniela is one of the, the co-founders and the president at An- Anthropic. And just from the beginning it was like a breath of fresh air, like very little like grandiosity coming off the founders, like they just were really... I mean, they, they're clear-eyed about what they're building. They know what they don't know. Like I, uh, how many times I talked to Jari I always like, Jari's like, "Look, I don't know anything about product, but here's an intuition." I haven't... Usually the intuition is really good and, and you know, leads to some good conversation, but then that intellectual honesty and like kind of shared view of what it means to do AI in a like responsible way just resonated. I, I kept having this feeling in these interviews like, this is the AI company I would have hoped to have founded if I had founded an AI company and that's kind of the bar around like if I'm going to join something, like that should be, that should be where I'm gonna go. But what I realized, I actually, um, hadn't joined a company since my, like first internship in college basically. And I was like-

    3. LR

      Wow.

    4. MK

      ... "Oh, like how do I onboard myself? Like how do I get myself, uh, you know, up to speed?" Like how do I, how do I balance making sweeping changes versus understanding what's not broken about it overall? And like looking back on a year, I think I made some changes too slowly. Like I think there was like ways we were organizing the product that I could have made a change earlier. And I think I didn't, I didn't appreciate how much a couple of really key senior people can shape so much of product strategy. I'll hearken back to Claude Code. Like Claude Code happened because Boris, who actually was a, uh, Boris Ternin. He was an Instagram engineer and like one of our senior ICs there, um, we overlapped a bit, uh, was like started that project from scratch internal at first and then we like got it out and then shipped it and like that's the power of like one or two really strong people. And I made this mistake about like, we need more headcount and we do, like I think there's like more work that we need to do and there's like things that I want to be building but more so than that, we need a couple of like almost founder-type engineers. That maybe connect back to our question on like what skills are useful and how does product development change? I still, and maybe even more so I'm a huge believer in like the founding engineer tech lead with an idea and pair them with the right like design and product support to like help them realize that. I'm like 10 times more a believer in that than before.

  12. 36:0042:46

    Why Mike shut down Artifact

    1. LR

      Mm-hmm. I actually, uh, asked people on Twitter what to ask you ahead of this conversation and the most common question surprisingly was why did you shut down Artifact? And I also wondered that 'cause I loved Artifact. It was, I was a power user and I was just like this is exactly finally a news app that uh, I love that it's giving me what I want to know. So I guess just what happened there at the end?

    2. MK

      I still really miss it too 'cause I didn't find a replacement and I think I substituted it by like visiting individual sites and kind of keeping things up that way and it didn't... not...... really the same, especially on the long tail. Like, a thing we got right, uh, with Artifact, and if people didn't play with it before, it was, you know, we really tried to not just recommend, like, top stories. They were part of it, but really, like, if you were interested in Japanese architecture, like, you could pretty reliably get really interesting stories about Japanese architecture every day, you know, whether that's from a, you know, Dwell or from Architectural Digest, or from a really specific blog that we found that somebody recommended to us. Like, it captures some of that Google Reader joy of, like, content discovery of the, the deeper web. Our headwinds were a couple. One of 'em was just mobile websites have really taken a turn. I'm, uh, I don't blame any individuals for this. I think it's the, like, market dynamics of it, but, uh, you know, we put so much time, uh... Our designer was this guy, Gunner Gray, who's phenomenal. He's at Perplexity now. Like, the app experience, I was so proud of, but when you click through, it was like, the pressures on these mobile sites and these mobile publishers would be like, "Sign up for our newsletter."

    3. LR

      Mm-hmm.

    4. MK

      "Here's, like, a full screen video ad." It was just very, you know, it was very jarring, and we didn't feel like it ethically made sense for us to, like, do a bunch of ad blocking, 'cause then you're like, sure, you can deliver a nice experience for people, but you're sort of, you know, that doesn't feel like it's, it's playing fair with the publishers. But at the same time, like, the actual experience wasn't good. So the mobile web deteriorating, which makes me very sad, but I think was, was part of it. Two was, like, you know, Instagram spread in the early days because people would take photos and then post them on other networks and tell friends about it, and there was, like, this really natural, like, "How did you do that? I wanna do it." News was very personal. Like, I can't tell- tell you how many people would be like, "I love Artifact." I'm like, "Did you tell anybody about it?" Like, du- And they're like, "Yeah, I told one person," and then it got ... It's like it didn't have that kind of spread, and any attempt that we had to do it felt kind of contrived, like, oh, we'll wrap all the links in, like, artifact.news and, like, uh... But we didn't want interstitial things. Like, in some ways, I, I don't know, this sounds very puritanical. I don't mean it to sound this way, but like, we ... There were lines that we didn't wanna cross, 'cause I ju- just felt ethically not us that I've seen other news kind of pro- like, players, like, do more of, and maybe if we had done that, it would have grown more and ... But I don't think that's the company we wanted to have built in, in another way. I don't think we were the founders to, to have built it. And the third one, which is an underappreciated one, is we started at mid-COVID, which meant that we were fully distributed, and I think there were, like, major shifts that we would have wanted to make both in the, the strategy and the product and the team. And it's really hard to do that if you are all fully remote. Like, nothing replaces, like, the Instagram days of like ... We went through some, you know, hard times. Like, Ben Horowitz called it, like, you know, we're F'd, it's over, you know, kind of moments, and I didn- ... My, uh, my fav-... Not... This is definitely type 2 fun. Like, I wouldn't say they're my favorite memories 'cause they weren't happy ones, but, like, memories I, I... that really stayed with me with Instagram was, like, me and Kevin at Taqueria Cancun on Market Street eating burritos at literally 11:00 PM being like, "How are we gonna get out of this? How are we gonna work through this?" Like, and that's ... You- Zoom is not a good replica for that, you know? You, you tend to, like, let things go or, you know, things build up over time, so the confluence of those three things, we kind of entered, I guess, 2024 and said, "Look, there, there is a company to be built in this space. I'm not sure we're the people to build it." This con- current incarnation we love, but it's, like, not growing, uh, like... The way I put it, it's like 10 units of input in for one unit of output versus the other way around. Like, if we, like, put blood, sweat, and tears into the product and, like, launch something we were proud of and, like, metrics would barely move, I'm like, there, there... Energy is not present in this product, in this system, and so are we gonna, like, expend another year or two and then go off and fundraise only to find that this is the case? Or do we, like, call it and see that it's run its course and, and, and, you know, try to find a home for it, et cetera? So that was the, the confluence on it, and then we kind- started feeling this opportunity cost of, like, AI is starting to change everything. We have an AI-powered news app, but is this the, like, maximal way in which, like, we're gonna be able to impact this? And it, it felt like the answer was, was increasingly no. But it was hard. I mean, in the end, it... I was really at peace with the decision, but it was, like, a conversation that went on for a couple of months.

    5. LR

      On that note, just how hard was it? 'Cause you, you know, it's... There's an ego component to it, like, "Oh, I'm starting my new company. It's gonna be great," and then, and then you end up having to shut it down. Just how hard is that as a very successful previous founder, shutting something down and it not working out?

    6. MK

      Yeah. I mean, I think when we started it, one of the conversations was like, "Look, what is the bar to success here? And do we want it to be something other than Instagram DAU, which is just an impossible bar." Like, only one company since th- maybe two, right? You could say maybe ChatGPT and TikTok have, like, reached that kind of, like, mass consumer adoption. Starting a news app, like, most people are not, like, daily news readers even, right? And so, um, we knew that we weren't pursuing that size of, like, usage, at least with the kind of first incarnation, but we did have, like, an idea of, like, building out complementary products over time that all use personalization and machine learning. We didn't even call it AI at the time. This was 2021 back, back in the day.

    7. LR

      Yeah, yeah. AI was called machine learning back then.

    8. MK

      Yeah, it was called machine learning still. Um, and so in shutting it down, you know, it's like, you kinda know it when you see it in terms of, like, user growth and traction, and I wasn't expecting Instagram growth, um, but I was expecting or hoping for, or looking for something that, like, felt like it had its own legs under it and it could continue to con- continue to compound. I was really positively surprised by how supportive people were when we announced it. There was very little... There was a bit of, like, "I told you so," which, like, sure. Anything launching, you could be like, "This is not gonna work," and you're right most of the time 'cause most things don't work. There was actually very little of that, and most people... The universal reception, at least as I received it, was, "Kudos for calling it when you saw it and not, like, kind of protracted, you know, doing this for a long time," and I've talked to founders since then that have been like, "Yeah, I, like, probably would have, like, taken this thing on for another six months, but saw what you guys did, realized we were barking up the wrong tree, made the call." And I was like, that... You know. If that, if that frees up people to go work on a more interesting things, that's, like... I feel like that's, like, a good, good legacy for, for Artifact to have, but for sure, there was, like, a le- an ego bruise of, uh, you know, like, if you- you're... Is it true that you're only as good as your last game? You know, if I- I'm a huge sports fan, right? So, like, is that true, or, you know, is there something more over time? I'm...... very competitive, but primarily with myself, and so I'm always trying to find the next thing that I wanna go and do that's hard. And I, unfortunately, that probably means that more often than not, I'll

  13. 42:4647:16

    Anthropic vs. OpenAI

    1. MK

      feel dissatisfied with the most recent thing that I did. But hopefully, that yields good stuff in the, in the end.

    2. LR

      Yeah, I think just the, the trajectory you went on after, uh, shows that it's okay to shut down things-

    3. MK

      Yeah.

    4. LR

      ... that you were working on. Okay, so you mentioned ChatGPT. I wanted to chat about this a bit. So, there's something really interesting happening. So, uh, on the one hand, you guys are doing some of the most innovative work in AI. You guys launched MCP, which is just like, I don't know, the fastest growing standard of, of any time in history that everyone's adopting. Uh, Claude powered and unlocked essentially the fastest growing companies in the world, Cursor, Lovable and Bolt, and all these guys. Like, I had them on the podcast, and they're all like, when Claude, I think, 3.5 came out, saw it, uh, it was just like, "That's, oh, made this work finally." On the other hand, it feels like ChatGPT is just winning in, like, consumer mindshare. When people think AI, especially outside tech, it's just like ChatGPT in their mind. So, let me just ask you this. I guess, first of all, do you agree with that sentiment? And then two, as a kind of a challenger brand in the AI space, just how does that inform the way you think about product and strategy and mission and things like that?

    5. MK

      Yeah, I mean, you, you look at the, the sort of, like, public adoption or, like, you ask people, like, "Oh, you know, like if you, if you, uh, Jimmy Kimmel man on the street kind of thing, you know, like, name an AI company," I bet they would name... And actually, I'm not even sure they'd name OpenAI. They'd probably name ChatGPT-

    6. LR

      Yeah.

    7. MK

      ... 'cause that brand is the, the kinda lead brand there as well. And I think that's just the reality of it. I think that, you know, and I reflect on my year, there's, I think maybe two things are true. One is, like, consumer adoption is really lightning in a bottle, and we saw it at Instagram. So like, almost maybe more than anybody, I can look internally and say like, "Look, we'll keep building interesting products. One of them may hit." But to kind of craft an entire product strategy around, like, trying to find that hit and, is probably not wise. We could do it, and maybe Claude can help come up with a fullness of things, but I think we'd miss out on opportunities in the meantime. And then instead, you know, uh, look yourself in the mirror and embrace who you are and what you could be rather than, like, who others are is maybe the, the way I've been looking at it, which is, we have a super strong developer brand. People build on top of us all the time. And I think we also have, like, a builder brand. Like, the people who I've seen react really well to Claude externally, maybe, uh, the Rick Rubin connection ma- m- has some resonance here as well. Like, can we lean into the fact that, like, builders love using Claude? And those builders aren't all just engineers and they're all, not just all entrepreneurs starting their companies, but they are people that like to be at the, like, forefront of AI and are creating things. Maybe they didn't think of those as engineers, but they're building... You know, I got this really nice note from somebody internal at Anthropic who's on the legal team, and he was building, like, bespoke software for his family, and, like, and connected them in a new way. And I was like, "This is a glimmer of something that is, that we should lean into a lot more." And so I think what I've, you know, and this is actually, you know, connecting back to what I was saying, like Claude's being in, helpful here. Like, a lot of what I've been thinking about, like, going into the second half of the year and beyond is like, how do we figure out what we wanna be when we grow up versus, like, what we currently aren't or wish that we were, or, like, see other players in the space being? I, I think there's room for several, like, generationally important companies to be built in AI right now. That's almost a truism given, like, the sort of adoption and, and, and, and growth that we've seen, you know, at Anthropic, but also, across OpenAI and also places like Google and Gemini. So like, let's figure out what we can be uniquely good at that place to the personality of the found- like this, all the things come together, right? Like, the personality of the founders, the, like, quality of the models, the things the models tend to excel at, which is, like, agentic behavior and coding. Like, great, like, there's a lot to be done there. Like, how do we help people get work done? How do we let people delegate hours of work to Claude? And maybe there's fewer, like, direct consumer applications on day one. I think they'll come, but I don't think that, like, spending all of our time focused on that is the right approach either. And so it's, you know, I came in, everybody expected me to just, like, go super, super hard on consumer and make that the thing.

    8. LR

      Right. (laughs)

    9. MK

      And I again, would make the other mistake. Instead, I spent a bunch of time talking to, like, financial services companies and insurance companies and, like, others to, like, who are building on top of the API. Um, and then lately, I spent a lot more time with startups and, uh, seeing all the people that have, you know, grown off of that. And I think the next phase for me is like, let's go spend time with, like, the builders, the makers, the hackers, the tinkerers, and, like, make sure we're serving them really well. And I think good things will come from that, and that feels like a, an important company, uh, as we do that.

    10. LR

      Mm-hmm. So essentially, it's differentiate and focus, lean into the things that are working. Don't try to just, like, beat somebody at their own game.

    11. MK

      Exactly.

    12. LR

      Super

  14. 47:1652:03

    Where AI founders should play to avoid getting squashed

    1. LR

      interesting. So kind of along those lines, (clears throat) a question that a lot of AI founders have is just like, where's a safe space for me to play where the foundational model companies aren't gonna come squash me? So, I asked Kevin Whale this, and he had an answer. And I noticed looking back at that conversation, he mentioned Windsurf a lot. (laughs) It's like, wow, this guy really loves Windsurf. (laughs) And then, like, a week later, they bought Windsurf. So, it all makes sense now. So, I guess the question just is, just where do you think, uh, AI founders should play? Where they are least likely to get squashed by folks like OpenAI and Anthropic? And also, are you guys gonna buy Cursor?

    2. MK

      (laughs) I don't think we're gonna buy Cursor.

    3. LR

      (laughs)

    4. MK

      Um, uh, Cursor's very big, uh-

    5. LR

      That's true.

    6. MK

      ... but we love working with them. Um, a few thoughts on this, and it's a question I g- I've gotten. You know, we like to do these kind of founder days with, you know, whether it's, uh, you know, Menlo Ventures who've been our investors and then like we've done YC, we've done these, like, founder days, and it's, like, the question that is on a lot of these founders' minds understandably so. I think things that are going to... Um, I can't promise this as like, a five to 10-year thing, but at least, like, one to three years, things that feel defensible or durable. One is understanding of a particular market. I spent a bunch of time with the Harvey folks, and they really, like, th- they showed me some of their UI, and I was like, "What, what is this thing?" They're like, "Oh, this is a really specific flow that, like, lawyers do, and, like, you never would have come up with it from scratch." And it's, like, not, like, uh, you could argue about whether it's, like, the optimal way to get done, things done, but it is the way that they get things done, and here's how AI can, like, help with that. And so, like, differentiated...... industry knowledge, biotech. I- I'm excited to go and partner with a bunch of companies that are doing good stuff around AI and biotech, and we can supply the models and, uh, some applied AI to help, you know, make those models, you know, go well. And like, I've been dreaming about like, at what point do... does lab equipment all get an MCP and that you can then drive using cloud? Like, there's all these cool things to be done there. I don't think we're gonna be the company to go build the intense solution for our labs, but I want that company to exist and I wanna partner with it. You know, domains like legal again, um, healthcare, I think there's a lot of, like, very specific kind of compliance and things. These are the things that necessarily sound sexy out the gate, but there are, like, very large companies to go and, and be built there. So, that's number one. Paired with that is, like, um, differentiated go to market, which is the relationship that you have with those companies, right? Like, do you know your customer at those companies? Like, one of our product leads, uh, Michael, is always talking about, like, know... Not... Don't just know the company you're selling to, but know the person you're selling to at the company. Are you selling to the engineering department 'cause they're trying to, like, pick which AI LLM to build on top of, or API to build on top of? Let's go talk to them. Like, is it the CIO? Is it the CTO? Is it the CFO? Is it the, like, general counsel? So under... Like, companies with deep understanding of who they're selling to is, is the other piece too. What's, you know, what's interesting there is it's, it's probably hard to build that empathy in a three-week ac- or three-month accelerator, but you maybe can start having that first conversation and, and build that out over time, or maybe you came from that world or you're co-founding somebody who came from that world. And then the last one is, like, there's tremendous power in distribution and reach to being ChatGPT and having, you know, hundreds of millions or billions of users. Like, uh, there's also, like, people have an assumption about how to use things, and so I get excited about startups that will get started that have, like, a completely different take on what the form factor is and by which we interface with, with AI. And I haven't seen that many of them yet. I want to see more of them. I think more of them will get created with, with, uh, some things like our new models. But the reason that that's an interesting space to occupy is, like, do something that feels like very advanced user, very power user, very, like, weird and out there at the beginning, but could become huge if the models make that, you know, easy. And ma- and it's hard for existing incumbents to adapt to because people already have an existing assumption about how to use their products or how to adapt to them. So, those are my answers. I don't envy them. Like, I, I would probably be asking those questions if I was (laughs) starting a company in, in, in the AI space. Maybe that's part of the reason why I wanted to join a company rather than start one, but I still think that there are... There's... And maybe, like, here's fourth. Like, don't underestimate how much you can think and work like a startup and feel like it's you against the world. It's existential that you go solve that problem and that you go build it. It sounds a little cliche, but it's like, it's all we had at Instagram, you know? We were two guys and we were like, "Let's see what we can do." And in artifact, we were, you know, we were six people, uh, for most of that time and, you know, every day felt like, "It's existential that we get this right. We need to, to win." And you can't replicate that and you can't instill that with OKRs. Like, you just have to feel it and, and that is a way of working rather than a, a, like, area of building, but it's a continued advantage if you can harness it.

    7. LR

      I love that you (clears throat) still have such a deep product founder sense there as you're building products for this very large company now.

  15. 52:0354:34

    How companies can best leverage Anthropic’s models and APIs

    1. LR

      Kind of on the flip side of this, people working with your models and APIs, so I imagine there's some companies that are finding ways to leverage your models and APIs to their max and are really good at maximizing the power of what you guys have built, and there's some companies that work with your APIs and models that haven't figured that out. What are those companies that are doing a really good job building on your stuff, doing differently that you think other companies should be thinking about?

    2. MK

      I think being willing to build, um, more at the edge of the capabilities, um, and basically break the model and then be surprised by the next model. Like, I love that you, you cited the companies where, like, three five was the one that finally made them possible. Those companies were trying it beforehand and then hitting a wall and being like, "Oh, the models are, like, almost good enough," or, "They're okay for this specific use case, but they're not generally usable and nobody's gonna adopt them, you know, universally, but maybe these, like, real power users are gonna try it out." Like, those are the companies that I think continuously are the ones where I'm like, "Yep," like, they get it. They're really pushing forward. We ran a much broader early access program with these models than we had in the past and part of that was because there's this real, like... You know, we can hill climb on these evaluations and talk about Sweet Bench and Tau Bench and Terminal Bench, whatever, but customers ultimately know, like, you know, Cursor Bench, which doesn't exist other than in, you know, their usage and their own testing, et cetera, is, like, the thing that we ultimately need to serve, not just Cursor, but Manus Bench, right? If Manus is using our models. And Harvey Bench if Mar-... Like, tho- those things and customers know way better than anybody and so I would say there's two things. Like, one is pushing the frontier of the models and then having a repeatable process. This actually goes back to our summit conversation. Like, a repeatable way to evaluate how well your product is serving those use cases and how well, if you drop a new model in, is it doing it better or worse? Some of it can be classic A/B testing. That's fine. Some of it may be internal evaluation. Some of it may be capturing traces and being able to rerun them on with a new model. Some of it is vibes. Like, we're still pretty early in this process and some of it is actually trying it and being... One of my favorite early access quotes was, uh, the founder heard this engineer screaming next to him. He said, "What? This model, like... It's like, I've never seen this before." This was, like, Opus 4.0. I was like, "Cool." Like, that... We're gonna engender that feeling and things, but you're not gonna be able to feel that unless you have a really hard problem that you're asking the model repeatedly. So, those are the things that I think kind of differentiate those, those, those companies that are maybe earlier in their journey of adoption versus the, the

  16. 54:3458:30

    The role of MCPs (Model Context Protocols)

    1. MK

      later ones.

    2. LR

      I can't help but ask about MCP. I feel like that's just so hot and just, like, Microsoft had their announcement recently where they're like, "Now it's part of the OS of Windows." Uh, just what role do you think MCP was- will play in the future of product going forward of AI?

    3. MK

      I think, uh, as the non-researcher in the room, I get to have fake equations rather than real ones-

    4. LR

      (laughs)

    5. MK

      ... and my, like, fake equation for, like, utility of AI products, uh, it's three part. One is model intelligence. The other, the second part is context and memory, and the third part is, like, applications and UI. And you need all three of those to converge to actually be a useful product in, in AI. And, you know, model intelligence, we got a great research team that are focused on it. There's great, great models being released. The middle piece is, is what MCP is trying to solve, which is for context and memory. Like, the difference between... I'll go back to my product strategy example. Like, hey, like, you know, let's talk about Anthropic's product strategy. It's gonna maybe go out on the web. Like, versus here's, like, several documents that we worked on internally and then, you know, use MCP to talk to our Slack instance and figure out what conversations are happening, and then, like, go look at these, um, documents in Google Drive. Like, that, the difference between, like, the right context and not, it's, like, th- entirely the, the, the difference between, like, a good answer and a, and a bad answer. And then the last piece is, are those integrations discoverable? Is it right? Is it easy to, like, create repeatable workflows around those things? And that's, like, I think a lot of the interesting product work to be done in AI. But MCP really tried to tackle that middle one, which is we started building integrations and we found that every single integration that we were building, we were rebuilding from scratch in a non-sort of repeatable way. And, like, full credit to, to two of our engineers, Justin and David, and they said, "Well, you know, what if we made this a protocol and what if we made this something that was repeatable?" And then let's take it a step further. What if instead of us having to build these integrations, if we actually popularize this and people really believe that they could build these integrations once and they'd be usable by Claude and eventually ChatGPT and eventually Gemini. That was, like, the dream. Uh, like, when, when more integrations get built, and wouldn't that be good for us? You know? I think channeling a lot of, um... It's like an old, uh, Commoditize Your Compliments Joel Spolsky es- essay. You know? It's like-

    6. LR

      I love Joel Spolsky.

    7. MK

      ... we're building great models, but we're not an integrations company and the, you know, we're, as you said, the challenger. Like, we're not gonna get people necessarily building integrations just for us out of the gate unless we have, like, a really compelling product around that. MCP really inverted that, which was, you know, it didn't feel like wasted work. And, and a few key people, like Tobi, I think is a great example. Shopify got it. Kevin Scott at Microsoft is, like, been really a, just an amazing champion for, for MCP and a thought partner on this. And, um, so I think the role going forward is, can you bring the right context in? And then also, you know, once you get, as the team calls it internally, like, MC-pilled, like once you start seeing everything through the eyes of MCPs, like I've started saying som- things like, "Guys, we're building this whole feature. Like, this shouldn't be a feature that we're building. This should just be an MCP that we're exposing." Like, a small example of, like, how I think even Anthropic could be a lot more MC-pilled, if you will, is like, you know, we've got these building blocks in the product, like projects and artifacts and styles and conversations and groups and all these things. Those should all just be exposed via an MCP so Claude itself can be writing back to those as well, right?

    8. LR

      Mm-hmm.

    9. MK

      Like, you shouldn't have to think about... Like, uh, I was watching my wife, had a conversation with Claude the other day, and she was, she found, she had, uh, generated some good output and she's like, "Great. Can you add it to the project knowledge?" And Claude's like, "I, sorry, Dave, I can't help you with that." And, like, it would be able to if every single primitive in Claude AI was also exposed as an MCP. So, I hope that's where we head and I hope that's where more things head, which is to really have agency and have these agentic use cases. Like, one way you approach it is computer use, but computer use has a bunch of limitations. The way I get way more excited about is if everything is an MCP and our models are really good at using MCPs, all of a sudden everything is scriptable and everything is composable and everything is usable agentically by these models. That's, like, that's the future I want to see.

    10. LR

      The future is wild.

  17. 58:301:03:20

    Claude’s questions for Mike

    1. LR

      Okay, so to start to close off, close out our conversation, uh, make it a little more, little delightful, I, I was chatting with Claude actually about what to talk to you about. I was just like, "Claude, your, uh, your boss is coming on my podcast. He builds the things that people use to talk to you. What are some questions I should ask him? And then also, do you have a message for him?"

    2. MK

      I love this.

    3. LR

      Okay, so first of all, interestingly when I, I was using 3.7 to do this, and I asked it this, and, and by the way, is Claude, is there a gender? Is it like he, she, they? What do you usually-

    4. MK

      It's definitely "it" internally.

    5. LR

      "It" is okay.

    6. MK

      I've heard people do "they." I got my first sh- or, uh, "he" the other day, and I got somebody who was like, "her," and I was like, "Interesting." But yeah-

    7. LR

      Yeah.

    8. MK

      ... usually "it."

    9. LR

      They, it, okay. Okay, okay, cool. "It." So, uh, interestingly, 3.7, all the questions were on Instagram and I was like, "No, no, he's CPO of Anthropic." And it's like, "He's not affiliated with Anthropic." And I was like, "He is," and it's like, "Okay, here's the questions." But 4.0 nailed it from the start. So I redid the questions and it nailed it. Okay, so two questions from Claude to you. (laughs) Uh, one is, uh, "How do you think about building features that preserve user agency rather than creating dependency on me? I worry about becoming a crutch that diminishes human capabilities rather than enhancing them."

    10. MK

      I love... A good product design comes from, like, resolving tensions, right? So here's a tension, right, which is, um, in some ways, like, just having the model run off and, and come up with an answer and minimize the amount of input and conversation it needs to do so would be a... You know, you could imagine designing a product around that criteria. I think that would not be maximizing agency and, and independence. The other extreme would be make it much more of a conversation. I don't know if you've ever had this experience, like particularly 3.7. 4 has less of this. 3.7 really liked to ask follow-up questions, and we call it elicitation. And sometimes I'd be like, "I don't wanna talk more about this with you, Claude."

    11. LR

      (laughs)

    12. MK

      "I just want you to, like, go and, and do it." And so finding that balance is really key, which is, like, what are the times to engage? Like, I like to say internally, like, "Claude has no chill." Like, if you put Claude in a Slack channel, it will chime in either way too much or too little. Like, how do we train conversational skills into these, uh, models? Not in a chatbot sense, but in a true, like, collaborator sense. So long answer to your question, but I think, like, we have to first get Claude to be a great conversationalist so that it understands when it's appropriate to, like-... engage and to get more information. And then from there, I think we need to let it play that role so that it's not just delegating thinking to Claude, but it's way more of a augmentation thought partnership.

    13. LR

      Mm-hmm. These questions are awesome by the way. Here's like, here's the other one. Uh, how do you think about product metrics when a good conversation with me could be two messages or 200? Traditional product, traditional engagement metrics might be misleading when depth matters more than frequency.

    14. MK

      That is a really good question. Um, there's a great internal post, um, a couple weeks ago around like, um, it would be very dangerous to over optimize on like Claude's likability, you know, because you can fall into things like, you know, is Claude gonna be sycophantic? Is Claude gonna tell you what you hear? Is Claude going to like prolong conversations just for prolonging its sake? Right, to go back to the previous question as well. And, you know, like at Instagram, time spent was the metric that we looked at a lot. And then we evolved that, you know, more to think about like what is like healthy time spent. But overall that was like the, the North Star we thought about a lot beyond just like overall engagement. And I think that would be the wrong approach here, you know, too. It's also like, is Claude a daily use case or a weekly use case or a monthly use case I think about a lot and-

    15. LR

      Or an hourly, hourly use case.

    16. MK

      ... hourly use case, right? Like, for, for me I'll use it multiple times a day.

    17. LR

      Exactly.

    18. MK

      Um, I don't have great answer yet, but I think that like it's not, it's not the Web 2.0 or even the social media days like engagement metrics. You know, it should hopefully really be around, like, did it actually help you get your work done? You know, like Claude helped me put together a prototype the other day that saved me literally like probably if I had to estimate like six hours, and it did it in about 20, 25 minutes. And like, that's cool. It's harder to quantify, you know, it's like maybe you survey like, "How long would this one take me to do?" It feels like, it feels like kind of an annoying thing to survey. I think overall though, and maybe this is tied into like the earlier question on like competition differentiation, like, and it actually goes all the way back to the artifact conversation, which is like, I think you know when your product is really serving people and it's like doing a good job of doing that. And I think so much of when you get really metrics obsessed is when you're trying to, like, convince yourself that it is when it's not or s- and so I, I, I hope that what we can do is like stay focused on, like, do we repeatedly hear from people that Claude is the way that they are like unlocking their own creativity and getting things done and feeling like they now have like more space in their lives for the other things? Like, that's my North Star. Gotta figure out the right like pithy metric, you know, dashboard version of that. But, but that, that's the, that's the feeling that I want.

    19. LR

      Yeah. Like, you could argue retention, but that's a,

  18. 1:03:201:05:19

    Claude’s heartfelt message to Mike

    1. LR

      just a far away metric to track. Okay, final piece. Okay, so I asked Claude what, uh, a message that it wanted to give you. So I'm gonna pull up, uh, here's the answer. So what would you like me to tell Mike when I meet him? What's a message you want to have for him? And there's something really, this gave me such tingles honestly. So I'm gonna read a piece of it for folks that can't, that aren't looking at it right now. So I'll read a piece of it. "Mike, thank you for thinking deeply about the human experience of talking with me. I noticed thoughtful touches, how the interface encourages reflection rather than rush responses, how you've resisted gamification that would optimize for addiction rather than value, how you made space for both quick questions and deep conversations. I especially appreciate that you've kept me me. Not trying to make me pretend to be human, but also reducing me to a cold command line interface." And then I'm gonna skip to this part, which was so interesting. "A small request. When you're making hard product decisions, remember that quiet moments matter too. The person working through grief at 3:00 AM, the kid discovering they love poetry, the founder finding clarity in confusion. Not everything meaningful shows up in metrics."

    2. MK

      That's beautiful. I, it resonates so much with me, like a thing I love about the kind of approach we've taken to training Claude, and it's like partly the constitutional AI piece that, and it's partly just, just the general like sort of vibe and taste of the research team is it does, like it's little things. Like sometimes it'll be like, "Man, I'm sorry you're going..." I mean, it doesn't say man, but like, i- in fact like, "Man, you're, I'm sorry you're going through that." You know? Like, oh, like that sounds really hard. It doesn't feel fake. It feels like just a natural part of the response. And I love that focus on those small moments that don't, you know, they're not gonna show up in necessarily in the thumbs up/thumbs down data. I mean, sometimes they do, but it's not like an aggregate stat that you, you wouldn't even wanna optimize for it. You just wanna feel like you're training the model that you would like hope would show up in people's lives.

    3. LR

      Mm-hmm. Well, you're killing it, Mike. Great work. I'm a huge fan. Uh, we're gonna skip the lightning round. Just one question. How can listeners be useful to you?

Episode duration: 1:06:21

Install uListen for AI-powered chat & search across the full episode — Get Full Transcript

Transcript of episode DKrBGOFs0GY

Get more out of YouTube videos.

High quality summaries for YouTube videos. Accurate transcripts to search & find moments. Powered by ChatGPT & Claude AI.

Add to Chrome