No PriorsNo Priors Ep. 70 | With Cartesia Co-Founders Karan Goel & Albert Gu
At a glance
WHAT IT’S REALLY ABOUT
Cartesia Bets on State Space Models for Real-Time Voice AI
- Cartesia co-founders Karan Goel and Albert Gu discuss their work on state space models (SSMs) like S4 and Mamba as efficient, elegant alternatives and complements to Transformers. They explain why SSMs are particularly well-suited for perceptual and multimodal data such as audio, and how this underpins their flagship product Sonic, a low-latency text-to-speech engine. The conversation covers technical trade-offs between SSMs and Transformers, hybrid architectures, and the potential to run powerful multimodal models on consumer devices instead of only in data centers. They also outline Cartesia’s roadmap toward multimodal conversational agents, on-device inference, and building a broader “rebellion” against Transformer-only thinking.
IDEAS WORTH REMEMBERING
5 ideasState space models offer linear-time sequence processing, making them ideal for long, streaming data.
Unlike Transformers’ quadratic scaling with sequence length, SSMs update a compressed state in constant time per token, which is crucial for audio, video, and other sensor data that require fast, continuous processing.
Transformers and SSMs are complementary, and hybrid architectures often outperform either alone.
SSMs act as ‘fuzzy compressors’ for bulk processing, while attention layers serve as exact retrieval or cache; many groups have found that mostly-SSM models with a small fraction of attention (around 10:1) work best.
Text-to-speech is far from “solved” when judged by human-level engagement and nuance.
Cartesia sees gaps in emotion, role-specific speaking styles, and long-form listenability; a practical test is whether you’d enjoy talking to the voice for more than 30 seconds.
High-quality speech systems increasingly require real language understanding, not just signal generation.
To pronounce words correctly and respond naturally, TTS and ASR need deeper semantic and contextual modeling, pushing systems toward integrated multimodal language models rather than isolated components.
On-device and edge AI will be a major next wave after large cloud models.
Cartesia is focused on making powerful models efficient enough to run in real time on laptops and eventually smaller hardware, reducing latency and cloud costs and enabling new applications that assume local intelligence.
WORDS WORTH SAVING
5 quotesPeople think you can throw a Transformer at anything and it just works. Actually, it doesn’t really.
— Albert Gu
I think of these state space models as fuzzy compressors, keeping a state in memory that’s always updating as you see new information.
— Albert Gu
The way I think about it is: would I want to talk to this thing for more than 30 seconds? If the answer is no, then it’s not solved.
— Karan Goel
The future will be more intelligence everywhere, and how do you enable that piece is what we’re excited about.
— Karan Goel
In the end, all the systems go away and it’s just one model.
— Albert Gu
High quality AI-generated summary created from speaker-labeled transcript.
Get more out of YouTube videos.
High quality summaries for YouTube videos. Accurate transcripts to search & find moments. Powered by ChatGPT & Claude AI.
Add to Chrome