World Models Are the Next Shift, Not Larger LLMs

dehakuran.com · May 2026 · 3 min read

Most people think today's AI revolution is about larger LLMs. The bigger shift is probably about something else entirely.


The frontier story of 2024–2025 was scale: more parameters, more tokens, more compute. The frontier story of 2026 is starting to look different.

The next-generation systems being funded today are not trying to predict the next word. They are trying to predict the next moment.

LLMs Learn Language. World Models Learn Reality.

The distinction is simple, and it matters.

  • LLMs learn the statistical structure of language — what word comes next.
  • World models learn the structure of reality — what happens next.

An LLM trained on the entire internet still has no internal model of physics, time, cause and effect, or persistent objects. It is extraordinary at producing plausible text. It is structurally unable to anticipate a falling glass.

That is what world models are designed to do.

Why the Shift Matters

The next generation of useful AI will not be limited to chat windows. It will need to:

  • Reason about dynamic environments, not static prompts.
  • Anticipate consequences of its own actions.
  • Adapt in real time to noisy, partial inputs.
  • Interact with the physical and digital world autonomously — robots, agents in real workflows, vehicles, medical-device assistants.

None of those are fundamentally text-prediction problems. They are reality-prediction problems.

"We're moving from 'What word comes next?' to 'What happens next?'"

The Efficiency Surprise

The most interesting part of the world-models story is that they are not chasing more compute. They are chasing less.

In March 2026, Yann LeCun's new venture AMI Labs (Advanced Machine Intelligence) raised $1.03B to build world models — after LeCun left Meta in late 2025 to bet his career on the thesis that LLMs are not the path to general intelligence.

The first public result, LeWorldModel (built with NYU, MILA, and Brown), is striking:

  • About 15 million parameters — orders of magnitude smaller than a frontier LLM.
  • Trains on a single GPU in hours.
  • Plans up to 48× faster than prior world models, on the tasks it was built for.

A research lab that does not need a hyperscaler-sized cluster to do interesting work is a different kind of competitive landscape. If "scale is all you need" was the LLM thesis, "structure is what scale was a proxy for" may be the world-model one.

Closer to AGI Than the Next Chatbot

This is the part I find most interesting.

Every additional parameter on a language model makes it a better text predictor. It does not, on its own, make it a better reasoner about the physical world. World models attack the harder, slower-moving problem: building systems that have an actual internal representation of what is going on.

The honest read is that LLMs have accelerated this work — the engineering, the funding, the talent flywheel — and any path toward AGI is going to braid multiple paradigms. World models are one of the more credible threads in that braid.

The next decade of AI may not be defined by the next-token race. It may be defined by which systems can answer what happens next — and do it on a single GPU.


Sources: AMI Labs (amilabs.xyz), TechCrunch on AMI Labs' $1.03B raise (March 2026), and the LeWorldModel research from NYU, MILA, and Brown. Views are my own and do not represent any commercial entity.

Frequently Asked Questions

What is a world model and how is it different from an LLM?

An LLM learns the statistical structure of language and predicts the next token. A world model learns the structure of reality — physics, causality, dynamics — and predicts what happens next. LLMs answer "what word comes next?"; world models answer "what happens next?" The first is sufficient for text generation; the second is what robots, autonomous vehicles, and any AI interacting with the physical world require.

What is AMI Labs and why does it matter?

AMI Labs (Advanced Machine Intelligence) is the startup Yann LeCun founded after leaving Meta in late 2025, raising $1.03B in March 2026 to build world models. It matters because LeCun has publicly argued for years that LLMs are not the path to general intelligence. AMI Labs is the well-funded bet that world models — not larger language models — define the next decade of AI.

Are world models more efficient than LLMs?

For their target tasks, yes. LeWorldModel — the first system from AMI Labs and collaborators at NYU, MILA, and Brown — runs on a single GPU with roughly 15 million parameters, trains in hours, and plans up to 48× faster than prior world models. That challenges the "scale is all you need" thesis: for predicting reality rather than text, smaller, smarter architectures may win.

AI StrategyWorld ModelsAGI

Deha Kuran

AI Executive, Engineer, and Evangelist. Head of AI Business Operations at Philips.

Follow the thinking on LinkedIn →