· Nolwen Brosson · Blog · 6 min read
JEPA vs LLM: Why Yann LeCun Thinks Generative AI Is a Dead End
For the past two years, the AI debate has often felt like a one-way conversation. On one side, LLMs take up all the space. On the other, a few researchers keep reminding us that aligning billions of tokens may not be enough to build a machine that truly understands the world.
Yann LeCun is one of them.
Meta’s Chief AI Scientist has been repeating the same idea for years: LLMs are useful and impressive, but they are not a credible path toward human-level intelligence. In his view, the future does not lie in models that generate one word after another, but in systems capable of predicting abstract representations of the world, anticipating, and reasoning. That is where JEPA comes in, short for Joint-Embedding Predictive Architecture.
So, is generative AI a dead end?
If the goal is to move beyond the competent chatbot and toward an AI that understands, predicts, and acts, then LeCun’s critique deserves more than a viral tweet.
JEPA vs LLM: Two Very Different Ways of Learning
An LLM mainly learns to predict the next token. This is a mechanism that makes it possible to produce coherent text, summarize, write code, and more. But the logic still relies on the statistical continuity of a sequence. The model excels at extending a form. It does not necessarily build a robust representation of the physical world, causality, or real-world constraints. That is exactly what LeCun is criticizing when he says that LLMs are useful, but not a path toward human-level AI.
JEPA follows a different logic. Instead of generating pixels, words, or frames one by one, the architecture tries to predict an abstract representation of a missing part from the observed context. This matters. The model is no longer asked to reconstruct the entire surface of the signal, but to capture what matters semantically. In his reference paper, LeCun presents JEPA and H-JEPA as non-generative architectures designed to learn predictive models of the world, with a hierarchy of representations.
Put differently:
- an LLM mainly learns to continue
- JEPA tries to learn what is plausible
- a World Model aims to predict what could happen next in the real world
This is a difference in philosophy before it is a difference in architecture.
Why Yann LeCun Criticizes LLMs
LeCun’s criticism is not that LLMs “do not work.” That would be absurd. They work very well across a wide range of tasks. His critique is deeper: they do not possess certain qualities considered necessary for a more general form of intelligence.
LLMs Model Language Well, Not Necessarily the World
Language compresses part of our knowledge, but only part of it. A child learns enormously by observing, manipulating, and testing hypotheses. LeCun emphasizes this point in A Path Towards Autonomous Machine Intelligence: humans and animals learn internal world models, meaning models of how the world works, largely through observation, and then use them to predict, reason, and plan.
LLMs Are Expensive, Fragile, and Often Superficial
An LLM can produce a brilliant answer, then fail on a question that requires respecting physical constraints, maintaining a coherent plan, or distinguishing what is plausible from what is merely grammatically likely. That conclusion is partly an inference, but it is consistent with the research agenda LeCun has defended since 2022.
JEPA, the Logical Successor to World Models
To understand why JEPA is generating so much interest, we need to go back to the idea of a World Model.
A World Model is a system that learns an internal representation of the world rich enough to anticipate how it may evolve, estimate what is probable, rule out what is impossible, and support decision-making. In his foundational text, LeCun presents world models as one of the main paths toward common sense and planning under uncertainty.
Why JEPA Is Closer to Reality Than Classical Generative AI
Traditional generative models often pay a high price to reconstruct every detail of a signal. But not every detail is useful. A shadow, visual noise, or a tiny texture variation does not always help with decision-making.
JEPA, by contrast, tries to learn more stable latent variables. The model predicts the representation of a missing region from another observed region. This choice forces the system to focus on meaningful structures, not on faithfully copying the surface. That is the core of Meta’s I-JEPA approach: a non-generative self-supervised learning method that predicts semantic representations rather than pixels.
From I-JEPA to V-JEPA 2
Meta has gradually expanded its research on JEPA. After I-JEPA for images, Meta introduced V-JEPA 2 as a world model learned from video, capable of understanding, prediction, and even planning in physical settings, with zero-shot robot control demonstrations in new environments.
In other words, this is no longer just an elegant theory. We are starting to see a concrete attempt to connect perception, anticipation, and action.
To go deeper on the topic from Fenxi’s perspective, you can also read this comparative analysis: World Models vs LLM.
Predictive AI vs Generative AI: Do We Really Have to Choose?
Framing “JEPA vs LLM” as a showdown is useful for grabbing attention, but less useful for understanding what is really happening. In practice, generative AI is not going away. It is too useful for interfaces, writing, coding, and content creation. The real question is this: what kind of architecture do we need to move beyond language utility and toward more robust intelligence?
The most serious answer today is probably this: LLMs will likely remain a major interaction layer, but they may not be the cognitive core of the most advanced systems. On that point, LeCun’s hypothesis is clear: the future lies more with predictive, hierarchical, and multimodal architectures than with simple next-token prediction.
Is Generative AI a Dead End?
The right answer is: very probably for strong AGI, no for short-term business.
LLMs have already found their market. They reduce the cost of producing text and accelerate software development. But if we believe the next revolution will come only from bigger models trained on more text, then LeCun’s critique is serious. His central argument is that an intelligence capable of understanding the world must learn hierarchical representations, handle uncertainty, anticipate the consequences of actions, and reason beyond surface-level correlations. That is exactly the territory of world models and JEPA-style architectures.
Conclusion: Is Yann LeCun Right?
On one point, probably yes: LLMs are not the whole of AI.
They dominate the headlines because they are visible, monetizable, and easy to demonstrate. But the next leap may come from architectures that are less spectacular in demos and more ambitious in depth: learning internal models of the world, abstracting what matters, predicting what may happen, and planning what to do next. That is exactly what JEPA is trying to bring to the table.
