Yann LeCun and Alex LeBrun joined Dr. Graham Walker to introduce AMI Labs and explain why they believe the next generation of AI in healthcare must move beyond large language models. While today’s LLMs are powerful and often surprisingly capable, Yann argues they are fundamentally designed to predict the next word — not to understand the world.
Large language models generate outputs by estimating probabilities. They predict what token is most likely to come next based on patterns in massive amounts of text. That approach has proven useful for summarization, documentation, and conversational interfaces. But in medicine, the stakes are different. Clinical decisions require reliability and the ability to reason about consequences. As Yann explains, an LLM produces actions “blindly,” without simulating what will happen next. It can appear intelligent, but it is not modeling cause and effect in the way physicians do.
Yann introduces the concept of “world models” as a different architectural approach. Instead of predicting tokens, a world model predicts how the state of the world changes in response to actions. In clinical terms, that means anticipating how a patient’s physiology might evolve after an intervention. During the conversation, Graham describes the moment when a physician pauses and mentally runs through possible treatment paths like fluids versus vasopressors or admission versus discharge. Yann responds that this is exactly the type of internal model AI would need to develop: a structured representation capable of simulating outcomes rather than generating text.
Alex adds a practical dimension from his experience building Nabla. In healthcare, accuracy thresholds are unforgiving. An AI system that is wrong even a small percentage of the time will quickly lose clinician trust. He describes how physicians will ignore a tool after encountering just a few incorrect outputs, much like they tune out unreliable EHR alerts. For AI to succeed beyond documentation, it must move from probabilistic suggestions to dependable reasoning systems. Yann argues that the real world is continuous, high-dimensional, and dynamic. It is far richer than the discrete tokens used in LLM training. World models, in contrast, aim to learn structured internal representations of those dynamics.
Alex and Yann emphasize that this shift will take time. Building generalized world models capable of understanding complex systems is still a research effort. But they envision a future where AI assistants function less like autocomplete engines and more like collaborative reasoning partners — systems that simulate, plan, and advise, while physicians remain in control. In that framing, the goal is not replacement, but amplification.
Thank you to our wonderful sponsor for supporting the podcast:
Sevaro is a physician-led telestroke and neurology company that delivers rapid virtual neuro coverage that’s reliable. Learn more at https://sevaro.com/
Here are the biggest takeaways for me about world models and their potential application to healthcare:
Join Offcall to keep reading and access exclusive resources for and by the medical community.
Offcall Team is the official Offcall account.
See what your colleagues are saying and add your opinion.