No PriorsNo Priors Ep. 29 | With Inceptive CEO Jakob Uszkoreit
At a glance
WHAT IT’S REALLY ABOUT
Transformer pioneer builds 'biological software' to reprogram life with RNA
- Jakob Uszkoreit, co-author of the Transformer paper and CEO of Inceptive, discusses the origins of the attention-based architecture and why its success is tightly coupled to modern accelerator hardware and community optimism. He argues that future AI progress must tackle elastic compute—models that dynamically adjust computation to problem difficulty and input complexity. Uszkoreit then outlines Inceptive’s vision of treating RNA as biological bytecode and medicines as compilable programs, using large-scale deep learning and custom assays instead of full mechanistic biological understanding. He suggests this black-box, end‑to‑end approach could dramatically expand the reach, scalability, and sophistication of medicines, especially mRNA-based therapeutics and vaccines.
IDEAS WORTH REMEMBERING
5 ideasArchitectures must be tightly matched to hardware to unlock breakthroughs.
The Transformer’s success came not only from the attention idea but from implementations that perfectly fit GPU accelerators, enabling massive parallelism and practical scaling compared to more sequential architectures.
Current models waste compute by not adapting effort to problem difficulty.
Today’s LLMs use computation roughly proportional to prompt and output length, not task hardness, leading to over-spending on trivial queries and under-spending on succinct but computationally hard problems.
Training on generated data can be valuable by amortizing prior compute.
Although synthetic data doesn’t add Shannon information, it can reuse past computational work; retraining on generated outputs effectively concentrates more compute on similar problems over time.
Elasticity in multimodal models is an underexploited efficiency frontier.
Models currently scale compute with input size (e.g., video length or resolution) rather than what is actually needed for the downstream task; more flexible architectures could adjust computation to information density and task complexity.
Deep learning enables powerful biology without full mechanistic understanding.
Uszkoreit argues that, as with language and many historical drugs, we can design effective biological interventions using data-driven, black-box models rather than waiting for complete, predictive theories of all underlying mechanisms.
WORDS WORTH SAVING
5 quotesAt the end of the day, the one thing we know really works in deep learning is making it faster and more efficient on given hardware.
— Jakob Uszkoreit
The big question is, does it matter that we may never test architectures that don’t fit today’s accelerators?
— Jakob Uszkoreit
Right now there’s no knob for a model to say, ‘This problem is hard, I should use more compute,’ versus ‘This is two plus two.’
— Jakob Uszkoreit
We think of RNA as the equivalent of bytecode and what we’re doing is compiling biological programs into RNA molecules.
— Jakob Uszkoreit
Maybe the hope to fully understand biology is actually holding us back; the ground truth is simply whether a treatment does more good than harm.
— Jakob Uszkoreit
High quality AI-generated summary created from speaker-labeled transcript.
Get more out of YouTube videos.
High quality summaries for YouTube videos. Accurate transcripts to search & find moments. Powered by ChatGPT & Claude AI.
Add to Chrome