Aakash GuptaAI Agents for PMs in 69 Minutes — Masterclass with IBM VP
At a glance
WHAT IT’S REALLY ABOUT
AI agents, RAG, open source, and the PM shift explained
- AI agents are positioned as the next leap beyond chatbots because they combine reasoning with planning, tool-taking actions, and reflection to automate end-to-end work.
- Building agents increasingly splits into two tracks: code-first frameworks (e.g., LangGraph, CrewAI, LlamaIndex, AutoGen) for control and low/no-code builders (e.g., Lindy, n8n, LangFlow, Flowise) for accessibility.
- RAG remains a dominant enterprise technique for injecting fresh, company-specific context into LLM workflows, but reliability requires serious data engineering and evaluation at multiple steps—not just checking the final answer.
- Managing “10–20 agents per employee” introduces a new orchestration skill: humans become accountable reviewers of agent outputs, with governance, cost controls, and safe experimentation as key enterprise constraints.
- AI changes product management by compressing the PM lifecycle (research → prioritization → PRD → prototype → monitoring) and enabling broader PM coverage, while still requiring customer-first problem investigation to avoid feature-factory behavior.
IDEAS WORTH REMEMBERING
5 ideasAgents matter because they close the loop from “answering” to “doing.”
Ruiz frames agents as the “wall of automation”: not just generating text, but decomposing tasks, executing actions in real systems (email/CRM/Workday), and improving through reflection over time.
The simplest useful mental model for agents is Think → Plan → Act → Reflect.
Thinking leverages LLM reasoning; planning breaks work into subtasks; action is enabled by tool access/protocols (he cites MCP); reflection uses feedback/history to iteratively improve future runs.
Pick your agent tooling based on required control, not hype.
Low/no-code builders accelerate experimentation for non-technical users, while code frameworks (LangGraph/CrewAI/etc.) remain necessary for complex, production-grade agentic systems needing deeper flexibility.
RAG is primarily for fresh context, not “making the model smarter.”
He distinguishes RAG (connecting to databases/knowledge bases for up-to-date info) from fine-tuning (better for behavior/style/specialization, not continuously changing enterprise knowledge).
Most RAG failures are evaluation and data-engineering failures, not “LLM failures.”
Enterprises can’t tolerate “70% accuracy,” so vanilla templates break; teams need systematic eval practices and better pipelines (embeddings, chunking, retrieval, filtering, ranking) to reach business-acceptable reliability.
WORDS WORTH SAVING
5 quotesAgents… deliver the wall of automation that is gonna unlock everyone… to generate way more output.
— Armand Ruiz
Four simple steps. The first one is thinking… planning… action… reflection.
— Armand Ruiz
70% accuracy is not acceptable.
— Armand Ruiz
Evals in agentic workflows should be almost… at every single step if you're really serious about developing something… a critical system.
— Armand Ruiz
If you didn’t write the most beautiful detailed PRD, still a lot of information is lost in translation… nothing speaks better than just a working prototype.
— Armand Ruiz
High quality AI-generated summary created from speaker-labeled transcript.
Get more out of YouTube videos.
High quality summaries for YouTube videos. Accurate transcripts to search & find moments. Powered by ChatGPT & Claude AI.
Add to Chrome