Skip to content
Aakash GuptaAakash Gupta

AI Product Metrics Interview – Execution Case Explained

We break down the complete AI product metrics framework. The 40-minute case walkthrough, visual framework approach, and why output metrics matter more than you think. Full Writeup: https://www.news.aakashg.com/p/ai-success-metrics-interview ---- Timestamps: 0:00 - Why AI Product Execution Interviews Matter 1:18 - The Underlord Case: Live Mock Interview Begins 4:29 - Why I Pulled Up The Product Live (Don't Skip This) 6:42 - The User Segmentation Push-Back 9:16 - Building The Visual Framework in Real-Time 12:52 - Value Enumeration: The 4 Core Values 17:24 - The Positive Metrics Bank 23:00 - North Star Selection: Why Exports Won 25:47 - Breaking Down The North Star (3 Vectors) 31:00 - Trade-offs & Guardrails Deep Dive 34:12 - The "Genie Metric" Curveball 35:01 - The Output Metrics Miss (My 8.5/10 Moment) 38:08 - The Power Move: Post-Interview Follow-Up Strategy ---- Key Takeaways: 1. Visual frameworks are non-negotiable - Draw your structure live. Interviewer needs to follow you for 40 minutes. Without visual anchor, even great ideas get lost. This separates you from every other candidate. 2. Always push back on user assumptions - Bart said "beginners." I challenged it. Underlord is on homepage, so metrics need to work for ALL users. This type of thinking turns 8/10 into 10/10 answers. 3. Build metrics bank BEFORE choosing North Star - Generate 15+ positive metrics first. Time, volume, adoption, discovery, engagement, retention, output. Then evaluate. Don't pick your favorite and work backwards. 4. Control for complexity in time metrics - Underlord might INCREASE time to edit because users do more. Create table: 1 tool = 3min, 2 tools = 4min. If Underlord takes longer for SAME complexity, that's a problem. 5. The output metrics mistake cost me 2 points - I forgot upgrades, renewals, referrals. Bart had to prompt with "genie metric" question. Always include input metrics AND output metrics. Non-negotiable. 6. North Star selection needs explicit reasoning - Don't just pick. Evaluate out loud. "Time to export doesn't work because... Number of tools is hard to operationalize because... Number of exports works because..." Show your math. 7. AI guardrails are different from traditional metrics - Hallucination rate. Support requests: no increase. Build evals that verify AI actually did what it claimed. This is table-stakes. 8. Break down your North Star by segments - New users vs power editors. Free vs paid. Short-form vs long-form. By equation: sessions × completion rate × exports per session. Makes it operationalizable. 9. The eval-driven approach for discovering failures - Ship to 10%. Collect traces. Review failure modes weekly. Create synthetic evals. Add new guardrails. This is the Hamel Husain / Shreya Shankar methodology. 10. The 30-minute follow-up wins jobs - Take your framework. Find gaps. Fix them. Mock up dashboard. Email interviewer. At Descript, you'll be the ONLY person who does this. Immediate differentiation. ---- 👨‍💻 Where to find Dr. Bart Jaworski: LinkedIn: https://www.linkedin.com/in/drbartpm/ Land PM Job: https://www.landpmjob.com 👨‍💻 Where to find Aakash: Twitter: https://www.x.com/aakashg0 LinkedIn: https://www.linkedin.com/in/aagupta/ Newsletter: https://www.news.aakashg.com #productmetrics #pminterview ---- 🧠 About Product Growth: Aakash Gupta's newsletter with over 200K+ subscribers. 🔔 Subscribe and turn on notifications to get more videos like this.

Aakash GuptahostDr. Bart Jaworskiguest
Jan 29, 202640mWatch on YouTube ↗

At a glance

WHAT IT’S REALLY ABOUT

Mock AI execution interview: choose North Star, guardrails, follow-up strategy

  1. Aakash demonstrates a repeatable success-metrics framework: clarify product/users, enumerate value, build a metrics bank, pick a North Star, decompose it, and add trade-offs/guardrails.
  2. The case centers on Descript’s Underlord, a natural-language AI agent that can access all editing tools, so success measurement must work for both novices and expert editors.
  3. The chosen North Star is “number of exports/publishes in 7–30 days,” justified as an end-to-end proxy for user value across time-saved, more output, and first-edit completion.
  4. Guardrails emphasize AI-specific risks—hallucinations, increased time-to-edit, user “rage interactions,” and support ticket volume—paired with an eval-driven approach using production and synthetic data.
  5. A key learning moment is the missed “output/business metrics” (upgrades, renewals, referrals), plus a “power move” post-interview follow-up: refine the dashboard afterward and email the interviewer with improved metrics and mockups.

IDEAS WORTH REMEMBERING

5 ideas

Start by validating the product’s actual capabilities—live—before proposing metrics.

Aakash pulls up the product to confirm Underlord is chat-based and has access to all Descript tools, ensuring the metrics map to real user actions and failure modes.

Define success from user value first, then translate value into measurable signals.

He enumerates four core values (faster editing, more edits/exports, first edit completion, publish/write-up assistance) and uses them to seed a coherent metrics bank.

Pick a North Star that reflects end-to-end outcomes, not isolated feature usage.

“Number of exports/publishes in 7–30 days” is chosen because it captures whether people actually finish and ship content, spanning both new and expert users.

Decompose the North Star along multiple vectors to diagnose where success or failure comes from.

He breaks exports down by (1) user type (new vs power), (2) export type (short vs long form), and (3) the underlying equation/action path (publish/export events).

AI products require explicit guardrails for quality, trust, and user friction.

He proposes guardrails like hallucination rate (<1%), time-to-edit not increasing (especially controlling for tools used), fewer support requests, and “accept with minimal edits” proxies.

WORDS WORTH SAVING

5 quotes

Today, we are giving you the very first full mock interview on YouTube ever published for AI product execution and AI product success metrics.

Aakash Gupta

We built it as a natural language alternative to old style editing... why not do essential thing like video editing... with your common words.

Dr. Bart Jaworski

Because Underlord is on the homepage, I really feel like the success metrics we need to have need to accommodate any user.

Aakash Gupta

What we would want is like these rage interactions with the chat.

Aakash Gupta

I probably should have included... some output metrics... upgrading plan... renewing... referring more people.

Aakash Gupta

AI product execution vs product sense interviewsUnderlord as an agentic natural-language editorValue enumeration tied to metricsPositive metrics bank and dashboard designNorth Star metric selection and decomposition vectorsAI guardrails, hallucination evals, and safety/privacyGenie metric and business/output metricsInterview communication: visual frameworks and follow-up

High quality AI-generated summary created from speaker-labeled transcript.

Get more out of YouTube videos.

High quality summaries for YouTube videos. Accurate transcripts to search & find moments. Powered by ChatGPT & Claude AI.

Add to Chrome