JEPAwiki

JEPA vs Alternatives

JEPA exists in a landscape of self-supervised and generative approaches. This page provides precise, mechanistic comparisons — not just "JEPA good, others bad," but what exactly differs in the computation.

The big picture: why this debate matters

Saining Xie, who has worked on both generative models (DiT/Sora architecture) and self-supervised learning (MAE, REPA, NEPA), frames the stakes bluntly:

"LLMs will never die, but will eventually fade. Old soldiers never die, they just fade away... It's a very good tool. I use LLMs every day. But it's not the foundation for building a general intelligence system."

"Language is a communication tool. Language is not a thinking map. Language is not even a decision-making tool."

"Your Language Model will gradually degrade to a simple communication interface, unlike now where all this multimodal intelligence is driven by large language models."

He argues LLMs are actually anti-Bitter Lesson: "Language is an extremely clever product of humans. It has intricate design." True self-supervised learning should work from raw observation (video, sensor data), not from humanity's most engineered communication protocol.

Rich Sutton's observation, cited by Saining: "Building the intelligence of a squirrel is the hard problem. Once you have a squirrel's intelligence... writing code, going to Mars — those things would be the easy ones." JEPA is the bet on building that squirrel-level understanding of the physical world.

JEPA vs Autoregressive LLMs

The dominant AI paradigm (GPT, LLaMA, Gemini) is autoregressive next-token prediction. JEPA rejects this for world modeling. Here's why, mechanistically.

What each predicts

An autoregressive LLM predicts:

p(token_{t+1} | token_1, ..., token_t)

A probability distribution over a discrete vocabulary. Every possible next token must receive probability mass.

JEPA predicts:

z_pred = f(z_context)

A single continuous vector in a learned representation space. No distribution, no vocabulary, no normalization.

The capacity waste argument

When an LLM predicts "The cat sat on the ___", it must assign probability to "mat," "chair," "roof," "table," "floor" — all valid continuations. The world-model-relevant fact (cats sit on surfaces) is diluted across thousands of plausible tokens. Most of the model's capacity is spent modeling surface-level variation (word choice, style, formatting) rather than world structure.

JEPA's target encoder compresses the target into a representation where "cat on mat" and "cat on chair" map to nearby points — both are "cat on surface." The predictor only needs to get the abstract structure right, not the exact surface form.

Sequential vs parallel

Autoregressive generation is inherently sequential: each token depends on all previous tokens. Generating N tokens requires N forward passes (or N decoding steps).

JEPA predicts all masked regions simultaneously. Predicting 5 future latent states costs the same as predicting 1 — they're just 5 rows in the same attention matrix. This isn't just faster; it prevents the error accumulation that plagues autoregressive rollouts.

World model: explicit vs implicit

An LLM's "world model" is implicit in its weights — there's no separate module you can query about the consequences of actions. You can only extract knowledge by generating text.

JEPA's world model is the predictor network itself. You can feed it a state and action, get a predicted next state, feed that back in, and roll out futures. This explicit world model enables planning via CEM — something no autoregressive LLM does from self-supervised pretraining alone.

What each is good at

Capability Autoregressive LLMs JEPA
Language generation Excellent Cannot
In-context learning Excellent Not demonstrated
Zero-shot task following Strong (via prompts) Not applicable
World modeling Weak / implicit Strong / explicit
Planning Via chain-of-thought (slow) Via latent rollout (fast)
Robot control Not from pretraining Yes (V-JEPA 2-AC)
Video understanding Good (with fine-tuning) Strong (from pretraining)

JEPA vs Diffusion Models

Both JEPA and diffusion models (Stable Diffusion, Sora, Cosmos) can be said to "predict in latent space." But the latent spaces serve fundamentally different purposes.

The latent space difference

Diffusion latents (from a VAE):

  • Must be invertible — a decoder must reconstruct every pixel
  • Must preserve all information, including noise, texture, lighting
  • Exist to make generation cheaper, not more meaningful
  • Are "compressed pixels," not "abstract state"

JEPA latents (from the target encoder):

  • Are deliberately non-invertible — many different inputs map to the same representation
  • Discard irrelevant detail by design
  • Exist to capture predictive structure, not visual fidelity
  • Are "world state," not "compressed data"

Concrete example: ball rolling off a table

Diffusion cares about: texture of the ball, lighting, background, pixel-perfect realism

JEPA cares about: the ball's position, velocity, the fact it will fall due to gravity

Only one of these supports reasoning.

Why diffusion cannot plan

Diffusion generates by iterative denoising — hundreds to thousands of steps, each a small local correction. This is:

  • Slow: generating one future frame takes seconds to minutes
  • Non-compositional: you can't chain denoised outputs to simulate long trajectories
  • Not action-conditioned: standard diffusion doesn't know about actions

V-JEPA 2-AC plans in 16 seconds per action. Cosmos (video generation–based planning) takes 4 minutes. LeWorldModel plans in under 1 second.

Can the two paradigms combine?

Yes. The REPA paper (arXiv:2410.06940) showed that aligning diffusion model internal states to pretrained vision encoder representations (like DINOv2) dramatically speeds up training (17.5x) and improves generation quality. This is JEPA-inspired thinking applied inside a generative model — use good representations to guide generation, rather than expecting generation to discover good representations.

A future system might use:

  • JEPA → world model & planning
  • Diffusion → rendering predictions into sensory detail

JEPA vs DINO / Contrastive Methods

DINO and DINOv2 are the closest relatives to JEPA, sharing the EMA teacher-student setup. But the learning signal is fundamentally different.

Invariance vs prediction

DINO learns invariance: "these two augmented views of the same image should have the same representation." The model learns what stays the same under transformations.

JEPA learns prediction: "given this visible context, predict the representation of the missing region." The model learns what must be true about unseen parts.

DINO:   z_student(crop_1) ≈ z_teacher(crop_2)
I-JEPA: z_predictor(visible) ≈ z_teacher(masked_region)

Why this matters for world models

Invariance is about identity (these views show the same thing). Prediction is about structure (this is what must exist given what I see). Only prediction extends naturally to:

  • Temporal dynamics (what happens next?)
  • Action conditioning (what happens if I act?)
  • Planning (which future is best?)

DINO representations cluster nicely and are great for classification. But they don't support latent rollouts because they were never trained to predict.

The augmentation question

DINO requires hand-crafted augmentations (crops, color jitter, blur, solarization) to define what should be invariant. These augmentations bake in assumptions — color invariance helps for classification but hurts for tasks where color matters.

JEPA uses masking instead. The only inductive bias is the masking strategy (what gets hidden), which is more general and doesn't inject domain-specific assumptions.

JEPA vs MAE (Masked Autoencoders)

MAE and I-JEPA look similar on the surface — both mask parts of an image and predict the missing parts. The critical difference is where prediction happens.

Mechanistic comparison

MAE pipeline:

visible patches → encoder → decoder → predicted PIXELS
Loss = ||predicted_pixels - actual_pixels||²

I-JEPA pipeline:

visible patches → encoder → predictor → predicted EMBEDDINGS
Loss = ||predicted_embeddings - teacher_embeddings||²

Why this matters more than it seems

MAE's decoder absorbs a lot of the complexity. The encoder doesn't need to produce maximally informative representations — the decoder can compensate. MAE can succeed by learning low-level texture interpolation without understanding objects.

I-JEPA has no decoder to hide behind. The encoder must carry all predictive structure because there's nothing downstream to compensate. Constant-vector encodings fail immediately because the predictor can't match varying teacher outputs.

Gradient pressure

In MAE, the encoder receives gradients through the decoder — diluted by the pixel-level loss. The decoder gets the strongest gradients.

In I-JEPA, the encoder receives direct gradient pressure to produce representations that are predictively useful. There's no intermediary.

This is why I-JEPA representations are:

  • More linearly separable
  • More transferable across tasks
  • More stable as world-model foundations

Representations are data — but at a different granularity

A common objection: "JEPA says it predicts representations, not data. But aren't representations data too?"

Yes — representations are data at a coarser, semantic scale. The distinction isn't ontological ("is it data?") but functional ("what information does it contain?").

Level Example Predictability
Raw data pixels, waveforms, tokens Mostly noise
Mid-level edges, patches, phonemes Partly predictable
JEPA representations objects, relations, state Predictable
High-level goals, intents Very predictable

The key property: JEPA representations are trained to be predictable. The target encoder discards information that doesn't help prediction. Many different inputs map to the same representation — that many-to-one mapping is the abstraction.

This mirrors classical state-space models: the observation x_t is raw sensor data, the state z_t is the sufficient statistics for prediction. Nobody in control theory confuses those two.

The energy-based perspective

LeCun often frames JEPA as an energy-based model. This doesn't mean there's a separate discriminator network. "Energy" is implicit in the prediction error:

E(context, candidate) = ||predictor(context) - candidate||²

Low energy = coherent, plausible future. High energy = implausible.

During planning, you minimize a combined energy:

E_total = E_world (JEPA consistency) + λ · E_goal (task objective)

The world energy keeps futures realistic. The goal energy pulls toward the desired outcome (e.g., "gripper is holding the object"). This means: reach the goal while remaining consistent with learned world dynamics — with consistency acting as a constraint, not a suggestion.

No discriminator. No adversary. No probability distribution. The predictor itself defines the energy landscape.

See also