Skip to content
How I AIHow I AI

Claude Code Just Got WAY More Powerful

I break down the biggest announcements from Anthropic’s “Code with Claude” event and what they actually mean for builders shipping AI products today. From scheduled AI routines to outcome-based agents, multi-agent orchestration, and new memory systems, I walks throug the features she’s most excited to use immediately—and how they could reshape the future of agentic software. *What you’ll learn:* 1. How Claude Code routines let you automate recurring workflows on schedules or webhooks 2. What “Outcomes” are and how rubric-based agent grading works 3. How multi-agent orchestration enables specialized AI teams with different roles and tools 4. Why Anthropic’s new “Dreams” memory system matters for long-term agent behavior 5. The biggest launch today (higher rate limits!) 6. How I think about building practical agentic products today *Links and resources:* • Code with Claude: https://claude.com/code-with-claude • Claude Code Routines Docs: https://code.claude.com/docs/en/routines • Define Outcomes Docs: https://platform.claude.com/docs/en/managed-agents/define-outcomes • Dreams Docs: https://platform.claude.com/docs/en/managed-agents/dreams • Multi-Agent Docs: https://platform.claude.com/docs/en/managed-agents/multi-agent • Managed Agent Webhooks Docs: https://platform.claude.com/docs/en/managed-agents/webhooks#supported-event-types • Codex (OpenAI): https://openai.com/codex • GitHub: https://github.com _Production and marketing by https://penname.co/._ _For inquiries about sponsoring the podcast, email jordan@penname.co._

Claire Vohost
May 7, 202611mWatch on YouTube ↗

EVERY SPOKEN WORD

  1. CV

    [upbeat music] Welcome back to How I AI. I'm Claire Vo, product leader and AI obsessive, here on a mission to help you build better with these new tools. Today, I attended Code with Claude, Anthropic's first developer event, and they announced some things in Claude Code and Claude Managed Agents I think you wanna know about. I'm gonna walk you through five things that launched today, how they work, what they are, and what I might build with them. We're gonna keep it under 10 minutes, and this is gonna be a quick preview of what you'll see in your new Claude Code and Claude API products experience. Okay. The first thing that shipped at Code with Claude that I think you'd wanna know about are some updates to the Claude Code app, and one of the updates that I know we've all been waiting for is routines, the ability to trigger events or actions on a schedule. We love it. You know I love OpenClaw, and what I love about OpenClaw are the crons. And so now Claude Code has that built in right here in the app. All you have to do is click new routine. You can either run it locally or remote. I'm gonna run it locally, and I'm gonna say weekly newsletter. This is something that I haven't been doing. I'm sorry if you're a ChatPRD customer. I know I haven't done my newsletter. I'm gonna say, "Look at our change log and draft a newsletter for us to send weekly." And then I'm gonna go in here and just say, "We have a changelog.md in the docs folder. Review it every Monday and write a customer-facing newsletter based on the best customer-facing features we shipped. Don't talk about behind the scenes things like tech debt or security unless they really impress customers." Okay, so I'm gonna do that. I'm gonna run it daily... Or no, sorry, I'm gonna run it weekly on Mondays at 6:00 A.M., and I think that's all I need to do. Oh, I'm gonna select my folder, um, where my project is, and then I'm gonna create, create. And now Claude Code will run my newsletter draft cron every week, and then I can come back in here, grab the HTML. If I were being really fancy, I would hook this up to my newsletter platform. I would hook it up to my Slack and ping us. But again, this is very useful to me. This is something that I used to kick off manually in Claude Code every Monday, and now I can do that here in Claude Code on a schedule. [lips smack] So how does that work? There are three trigger types. You can trigger them on a cron, which is a schedule, um, HTTP or a GitHub webhook. So you can do sort of a normal webhook or a GitHub webfoot. You can trigger these three ways, scheduled, like I just did, off of a GitHub action, or a general webhook. So you can hook it up to other systems to kick off a routine. All the stuff in connectors come along, so I have Slack connected, I have GitHub connected. So you can use those things as part of your routine, and it can run in the cloud, or it could run on your laptop, like I showed. And this is an example of a use case where you could say, "Weekly, I want you to check every PRD modified this week and check if it matches our rubric and post a summary to the team channel." So that's item one. [lips smack] The second one is in Claude Managed Agents in the API. If you haven't paid attention, OpenAI released something in Codex called Goal. You can do /goal in beta in Codex, and it'll basically bang its head against the problem, do what's called a Roth loop against a problem until it actually hits the goal. Anthropic released something very similar in the Claude API called Outcomes. You define what done looks like for an agent. It can self-grade and iterate until it gets there. There's a couple interesting things you need to know about how outcomes are defined. They all anchor on what's called a rubric, so there's a markdown file that's uploaded either through the files API or in line, and it's gonna tell your agent what success looks like. Then there is a grader, and it can do up to 20 iterations on the task to get to the outcome that you're going for. I wanna walk through this one in a very specific example to make this just a little bit more concrete for folks. So, so imagine that I want you to ship a ship-ready PRD, and I don't know if you can relate to this. Often you go through feedback cycles. You have to check it against priorities. You have to check it against technical capabilities. Now, using a Claude managed agent, you could, in theory, write a rubric, which is what does a good ship-ready, um, PRD look like? And then the agent can just take your PRD or your idea and iterate over and over and over again until it's fixed. Of course, you could expe- of course, you could expose this to your customers in an app, like I might do for ChatPRD. But again, this idea of outcome is define a rubric, give the agent the task, let it bang its head against that at least 20 times till it gets it right. I think this is a really interesting model for agentic products, and something I suspect many of us will use. The second thing I really love is a multi-agent framework supported in Claude Managed Agents. So now you can, through the API, explicitly define a multi-agent team that's going to work against the same container, the same file system, up to 25, which is kind of amazing. You can have, uh, uh, a orchestrator and then delegates, and so there's explicit hierarchy, and each agent can have its own tool set. I think this is really cool 'cause now you're able to define not just individual agents, but teams of agents programmatically through the API. And so the example I would give for something like ChatPRD is you could have a PRD orchestrator. This is sort of like the master agent that is intended to define and drive the work across the team. And then you could have three pieces or three sub-agents.A strategy agent that reflects the CPO voice, the critic agent that's sort of, like, supposed to poke at the holes in the PRD. I like being the critic agent. And then end review that could maybe have access to something like GitHub to optimize the technical implementation of the PRD. And so you can define this, as you see over here in the API. You define an agent in the API, you give it, um, a orchestrator level set of tools, and then you can define the sub-agents, um, in the, in the API with their own set of tools. And then you could expose that, as you can see here on the right, as three agents all working in parallel against the same problem owned by the coordinator or orchestrator level. Again, I think this is an interesting enhancement on the primitives of agents that people are gonna be using quite a bit. Okay, the next one I really like. It is Dreams. So this is all about agent memory. Just to make it simple for folks, memory is basically the idea of writing markdown files to the file system your agent uses that helps it do a better job the next time. It's not that fancy. Often those files have a date on them. But you don't really have to overthink it. But creating those memories is a little hard, and often a lot of the harnesses right now write memory on a hook. They write them on an event. And so what they do is, like, when you close a session, it writes memory, or when something happens, it write memory, or, like with OpenClaw, you can explicitly tell it to write memory. But what I like about Dreams, which is a very funny brand for an agentic memory product, but we'll allow it, is it's a primitive to call against a list of agent sessions. So let's say you've done 50 things with your agent. It's an explicit call to take those 50 sessions, review them, and then come up with important memories to write to disk. And as I'm saying this, I guess this is what we do when we dream. We go through our day, we review it silently, and then we decide what to commit to memory. I don't know. I don't know if this is the perfect metaphor, but it's the one we got, and it looks great on a, on a branded website. This one's in research preview, so I don't think everybody has access to this through API. I certainly don't have access to it, so I'm looking forward to touching it. But why I think this one's important to know is it just gives you a frame of reference for how Anthropic and these labs are thinking about the primitives, again, of agents and agent memory. And you can predict that some framework like this is going to be integrated into agentic platforms or agentic products where on some action or some revu- regular cadence, you're gonna review past sessions, and you're gonna explicitly write the right things to disk so they can be referred to moving forward. Side note, I think we think a lot about agent memory, but not a lot about agent forgetting, so I'm looking forward to, like, the purge version of this, which is dreams that tell you what to forget. I don't know if that's, like, trauma erasure or whatever, but [laughs] I think there's something interesting here. Okay, and then number five, the only announcement people really care about, which is usage limits are up. So starting today, Claude Code's five-hour limits are now doubled across Pro, Max, Team, and seat-based enterprise platforms. Peak hours are going away for Pro and Max plans, and the rate limits for Opus models in the API are going up. So we can all use these products more. Again, what did we see today at Code with Claude? Lots of other stuff. They might put data centers in space. There was a partnership with SpaceX. All sorts of fun things are happening. But the things that I think you can use immediately, routines for scheduling tasks in Claude Code, so you can get things done either on a webhook or on a schedule. Outcomes, which is the ability to set a rubric and task and have an agent work against that task at least 20 times to nail the rubric. Multi-agent orchestration, which allows you in the API to define an orchestrator role and sub-agents, I think up to 25, to get work done from different points of view with different tools. Dreams, which are a way to consolidate agent memory over sessions over time and do that on demand. And then finally, we can all use more Claude Code, which makes everyone happy. I do not know if these limit increases apply to Claude Design. I suspect they do not, but I hope they will. I think they launched a lot of stuff that we can use practically right now in Claude Code and gave me some good ideas for what the future of agentic products look like. None of this was mind-blowing, but I think it's all really practical, and it just gives you the idea that Anthropic is trying to be the agent platform of choice for builders. I look forward to seeing what you build with it, and I'm gonna go play with the API. Bye. [upbeat music] Thanks so much for watching. If you enjoyed the show, please like and subscribe here on YouTube, or even better, leave us a comment with your thoughts. You can also find this podcast on Apple Podcasts, Spotify, or your favorite podcast app. Please consider leaving us a rating and review, which will help others find the show. You can see all our episodes and learn more about the show at howiai pod.com. See you next time. [upbeat music]

Episode duration: 11:37

Install uListen for AI-powered chat & search across the full episode — Get Full Transcript

Transcript of episode efVfydaUIrM

Get more out of YouTube videos.

High quality summaries for YouTube videos. Accurate transcripts to search & find moments. Powered by ChatGPT & Claude AI.

Add to Chrome