Aakash GuptaThis AI Expert's Method Will Change How You Do Customer Research
CHAPTERS
Why AI user research feels unreliable—and the human-style fix
Aakash frames the problem: AI analysis can be messy and hallucinatory. Caitlin sets the core principle for making AI reliable—replicate rigorous human research workflows instead of jumping straight to “insights.”
The rigorous workflow: analysis first, then verification, then synthesis
Caitlin breaks down what “good analysis” means across interviews, surveys, and short feedback. The goal is to force structured, stepwise work that mirrors how researchers manually comb through raw data.
Tools and scope: surveys + interviews, plus automation vs. agentic workflows
They outline what will be demonstrated: a Claude-based workflow in the web UI, and an “agentic” parallelized version in Claude Code. Caitlin distinguishes common automation tools (Zapier/n8n) from more advanced agent setups.
AI-moderated interviews: promise, quality variance, and why craft still matters
Caitlin explains AI moderation tools and the wide quality range in their questioning and bias handling. Aakash underscores that even leading AI labs can ship mediocre interviewing, reinforcing the need to learn research craft.
Step 0 (Interviews): load context separately to prevent instruction overload
Caitlin demonstrates a dedicated “context-only” prompt before any analysis. This reduces model overload and ensures the LLM understands business goals, product tiers, and domain specifics—like onboarding a new teammate.
Why Claude for analysis + structuring transcripts with markdown to beat token limits
Caitlin explains her preference for Claude’s nuance, while noting Gemini’s strength in frequency/accuracy. She also shares a practical tactic: converting transcripts into structured markdown to improve both token handling and analytical accuracy.
Step 1 (Interviews): per-participant extraction—value anchors, fragile points, stability
Instead of asking for themes, Caitlin forces Claude to analyze each participant file in detail. The prompt extracts retention drivers and churn risks with explicit rating definitions and quote rules, producing a granular foundation for later synthesis.
Verification step (Interviews): contradiction checking to prevent cherry-picking
Caitlin runs a stress test to find contradictions within each participant’s statements and adjust risk ratings. This catches subtle inconsistencies that AI might otherwise blur when it tries to tell a coherent story.
Survey analysis setup: why you must code open-text responses before counting
Switching to churn survey data, Caitlin warns against immediately asking for top themes or frequency counts. She introduces coding (qualitative labeling) as the necessary first step to avoid miscategorization and missed nuance.
Survey Step 1: inductive open coding with rules + compute counts safely
Caitlin demonstrates a structured coding prompt: one primary mutually exclusive code per response, clear criteria, and an emphasis on building a usable codebook. She also recommends doing math via code (e.g., Python/pandas) to avoid LLM arithmetic errors.
Add emotional intensity ratings: beyond sentiment to prioritize what to fix
Rather than generic sentiment analysis (where churn feedback is “all negative”), Caitlin adds an intensity layer to distinguish soft exits from genuinely angry detractors. She calibrates the scale with definitions and few-shot examples including the reasoning behind labels.
Audit Step 3 (Surveys): AI checks its own coding and intensity for overstatement
Caitlin runs an audit prompt to recheck high-intensity ratings and code consistency, surfacing corrections that would otherwise undermine credibility. The output shows recoding and downgrading exaggerated intensity levels, making conclusions safer to present.
Claude Code agentic version: parallelize interview + survey analysis and export outputs
Caitlin shows how to encode the workflow into Claude Code agents (“Interview Analyzer Lite” and “Survey Analyzer Lite”) that pull from reusable context files and run analyses in parallel. The agent workflow outputs structured markdown reports with executive summaries, per-participant detail, quotes, and traceability.
Get more out of YouTube videos.
High quality summaries for YouTube videos. Accurate transcripts to search & find moments. Powered by ChatGPT & Claude AI.
Add to Chrome