MAR 17, 2026

What's Missing Between LLMs and AGI - Vishal Misra & Martin Casado

TRACK CAUSAL AIDEMYSTIFY TRANSFORMERSBUILD CONTINUOUS LEARNINGWATCH AGI MILESTONES

Key Takeaways

  • LLMs function through predictable mathematical updates - Experiments reveal that transformers refine their predictions in a precise, measurable way as they process data, rather than through inexplicable 'magic'.

  • AGI necessitates post-training learning - A critical gap in current models is their static nature; true AGI requires the ability to continuously acquire and integrate new information after the initial training phase.

  • Success depends on shifting from patterns to causality - Reaching human-level intelligence requires models to move beyond statistical pattern matching toward a fundamental understanding of cause and effect.

Episode Description

Vishal Misra returns to explain his latest research on how LLMs actually work under the hood. He walks through experiments showing that transformers update their predictions in a precise, mathematically predictable way as they process new information, explains why this still doesn't mean they're conscious, and describes what's actually required for AGI: the ability to keep learning after training and the move from pattern matching to understanding cause and effect.

Featured in Category Feeds

Stay in the Loop

Get AI + a16z summaries and more, delivered free.