How I AIThe power user’s guide to Codex | Alexander Embiricos (product lead)
CHAPTERS
Codex as a thorough “software engineering teammate” (setup + what it’s good at)
Claire introduces Alexander Embiricos (Codex product lead) and frames Codex as a coding agent optimized for diligence and complex tasks rather than raw speed. The episode’s goal is a practical, zero-to-one walkthrough from install to advanced workflows.
Installing the Codex VS Code extension and running a repo locally
Alex demonstrates the true beginner flow: install the extension, log in, and ask basic questions about an unfamiliar repo. The emphasis is that Codex is useful even before you write code—simply to understand how to run and interact with a codebase.
First fixes in a toy project: adjust jump + implement missing feature
With the game running locally, Alex uses plain-English prompts to fix a too-high jump and implement windmill planting. This showcases the basic “tell the agent what you want” loop and how Codex proposes a plan before editing code.
Parallel tasks: when to run multiple Codex chats vs serial work
Claire highlights parallelization as a key productivity lever, then asks how to decide between parallel and serial execution. Alex explains it depends on whether tasks are exploratory/questions (great in parallel) versus code changes that may conflict.
Git worktrees for safe parallel code changes (with Codex generating commands)
Alex introduces Git worktrees as the practical solution to parallel changes that must remain isolated. He demonstrates asking Codex to create two worktrees (French/German) so multiple variants can be developed concurrently without conflicts.
Terminal productivity: one-line prompts + working in multiple tabs
They zoom in on command-line efficiency: launching Codex with the first prompt inline and using multiple terminal tabs for parallel work. Claire notes many power users run many instances, but the key is structuring separation (e.g., via worktrees).
Case study: building the Sora Android app in 28 days with Codex
Claire asks how Codex scales to real production work; Alex cites OpenAI’s Sora Android app built by four engineers in 28 days and reaching #1 in the App Store. The takeaway is that agents don’t remove hard engineering—they compress timelines when paired with strong architecture and process.
Planning workflow with PLANS.md: “plan the work, then implement”
Alex demonstrates using a PLANS.md template (from an OpenAI blog post) to produce thorough, milestone-driven plans for large tasks (e.g., Python SDK mirroring a TypeScript SDK). He stresses iterating on the plan in the same chat to preserve context before execution.
Prototyping vs production engineering: vibe coding to learn, specs to ship
They distinguish two acceleration modes: rapid learning via prototypes and rapid execution for defined work. Designers can prototype Codex surfaces quickly, then engineering either lands it directly or rebuilds cleanly using the learnings and a tighter spec.
What needs a plan? Time, task complexity, and ‘Best of N’ exploration
Claire asks how to decide when to plan; Alex says it depends on task difficulty and personal time constraints. He describes “Best of N” (multiple attempts in parallel) as an alternative to up-front planning when speed matters or uncertainty is high.
Multiplying impact with code review: /review and GitHub automated reviews
Alex calls GitHub/code review the highest-leverage integration. He shows manual review via a /review command and automated GitHub PR comments that surface only high-confidence issues—then a loop where a developer can ask Codex to fix the issue directly.
Codex adoption at OpenAI: ‘everywhere,’ productivity shifts, and failed automations
Alex describes rapid internal adoption: moving from ~half the company to nearly all technical staff using Codex constantly. He shares that automated reviews are enabled across repos, while other proactive automations (e.g., auto-revising after human feedback) were less successful due to missing context and notification fatigue.
Harness matters: model-output quality + UX for latency, context, and iteration
Claire asks why the interface (“harness”) differentiates amid rapid model releases. Alex explains harness work improves both outcomes (tool usage patterns, compaction, parallel tool calling) and usability (seeing edits in IDE, handling latency, proactive helpfulness).
Atlas and personalized AI workflows: Sidechat, memory, and ‘be polite’
In a lightning-round segment, Alex shares how Atlas/ChatGPT becomes his default for non-code tasks due to personalization and memory. He highlights Sidechat for page-aware Q&A and rewriting, and argues for politeness to AI as a way to protect human habits and culture.
Getting agents back on track + closing: context, ambiguity, and session replay
Alex advises giving richer context and being explicit about ambiguity rather than over-specifying outcomes. If the agent derails, he recommends starting a new chat—and notes an advanced trick: Codex stores sessions locally, so you can have it read prior logs and continue coherently.
Get more out of YouTube videos.
High quality summaries for YouTube videos. Accurate transcripts to search & find moments. Powered by ChatGPT & Claude AI.
Add to Chrome