Aakash GuptaHow AI PMs Ship Features Users Love (Descript CEO Explains)
At a glance
WHAT IT’S REALLY ABOUT
Descript CEO on building AI editing tools and PM leadership
- Descript’s early AI feature strategy focused on packaging reliable, job-based “buttons” (e.g., remove retakes, edit for clarity) rooted in well-understood user workflows rather than novelty prompts.
- The team shipped AI tools using pragmatic, human-driven evaluation against real production data, iterating via public beta, adoption/retention metrics, and whether users exported the AI-modified output.
- As user needs became too parameter-heavy for fixed tools (e.g., Create Clips requesting endless knobs), Descript shifted toward Underlord, an objective-driven, open-ended co-editor agent.
- Underlord’s rollout emphasized tool coverage, representative regression tests, real-customer private alpha feedback, and improved activation—especially helping novices “get over the hump” of video editing.
- Burkhauser frames the PM’s unique value in AI as defining success/failure criteria for evals, while her career advice stresses deep product/customer command, shipping excellence, and humility in founder-led environments.
IDEAS WORTH REMEMBERING
5 ideasGreat AI features start with a concrete workflow pain, not the model.
Descript mapped creator workflows (scripted vs. improvised) and attached AI to specific pains like retakes, eye contact, and clarity—then hid prompts behind dependable, job-based buttons.
Ship “reliable buttons” first; use agents when customization explodes.
Fixed tools work well when inputs are bounded, but Create Clips requests kept adding parameters; Underlord emerged as the right abstraction once users needed highly customized, conversational control.
Human evals against real data are a valid starting point—if you’re disciplined.
Before formal eval stacks were common, the team tested on production-like content and shipped when results were genuinely usable; later they layered regression tests, A/B tweaks, and more automation.
PMs uniquely own the definition of “quality” for AI outputs.
Burkhauser argues only the PM can codify what “good,” “acceptable,” and “harmful” look like because it requires deep job/context understanding (e.g., judging jump-cut density, not just grammar).
Representative eval data matters more than sophisticated scoring.
Studio Sound quality regressed when evaluators used unrealistically terrible audio; the best model for ‘disaster audio’ differed from the best for the common ‘laptop mic’ use case, so the dataset must match the target workflow.
WORDS WORTH SAVING
5 quotesThe best products out there, they don't just do a job for you. They transform how you feel about yourself.
— Laura Burkhauser
Build them in these prepackaged, parameterized, job-based buttons that can give you a reliable result over and over again.
— Laura Burkhauser
You and only you are qualified to write the eval criteria for what… a good job looks like.
— Laura Burkhauser
What it didn't take into account is… how many jump cuts per 10 seconds are you putting into my video?
— Laura Burkhauser
If you're allowing for emergence, you're also allowing for a lot of, like, whack stuff to happen in your product.
— Laura Burkhauser
High quality AI-generated summary created from speaker-labeled transcript.
Get more out of YouTube videos.
High quality summaries for YouTube videos. Accurate transcripts to search & find moments. Powered by ChatGPT & Claude AI.
Add to Chrome