[LeWM] How to Train Stable World Models from Pixels with Just Two Loss Terms
Paper at a Glance
- Paper Title: LeWorldModel: Stable End-to-End Joint-Embedding Predictive Architecture from Pixels
- Authors: Lucas Maes, Quentin Le Lidec, Damien Scieur, Yann LeCun, Randall Balestriero
- Affiliation: Mila & Université de Montréal, New York University, Samsung SAIL, Brown University
- Published in: arXiv, 2026
- Link to Paper: https://arxiv.org/abs/2603.19312
- Project Page: https://arxiv.org/abs/2603.19312 (Code and Website references in paper)
The Gist of It: TL;DR
In one sentence: This paper introduces LeWorldModel (LeWM), a novel Joint-Embedding Predictive Architecture that learns stable world models directly from raw pixels by combining a simple prediction loss with a Gaussian-enforcing regularizer, eliminating the need for complex stabilization heuristics while enabling 48× faster planning than foundation-based alternatives.
Why It Matters: The Big Picture
For autonomous agents to navigate and interact with the real world, they need a “world model”—an internal simulation of how the environment responds to their actions. Historically, researchers built generative world models that tried to predict the future pixel-by-pixel. However, rendering every blade of grass or shadow is computationally exhausting and often irrelevant to the actual task (like pushing a block).
Enter the Joint Embedding Predictive Architecture (JEPA). Instead of predicting pixels, JEPAs compress observations into a low-dimensional “latent space” and predict future states in that compressed space. But JEPAs have a fatal flaw: representation collapse. Because the model only wants to minimize the difference between its prediction and the future state, it often cheats by mapping every observation to the exact same constant vector.
To prevent this “cheating,” prior works relied on a fragile cocktail of hacks: stop-gradients, exponential moving averages (EMAs), massive pre-trained frozen vision models (like DINOv2), or wildly complex loss functions with half a dozen tunable hyperparameters. LeWorldModel (LeWM) strips all this away. It proves that you can train an end-to-end JEPA from scratch on a single GPU using a mathematically principled, two-term objective function.
The Core Idea: How It Works
1. The Problem They’re Solving
Think of a JEPA model like a student taking a multiple-choice test where they write both the questions and the answers. If the goal is just to make the question match the answer, the easiest strategy (collapse) is to make the answer to every question “C”. Previous methods (like PLDM) fixed this by adding up to six different regularization penalties, creating a precarious balancing act that is notoriously difficult to tune. Foundation-model methods (like DINO-WM) sidestepped it entirely by freezing a pre-trained encoder, meaning the model couldn’t adapt its visual representations to specific task dynamics.
2. The Key Innovation
The authors fix this with a single, elegant mechanism: the Sketched-Isotropic-Gaussian Regularizer (SIGReg).
SIGReg forces the embeddings to spread out into the shape of a high-dimensional, isotropic Gaussian distribution. If the embeddings are forced to scatter across this bell-curve shape, they physically cannot collapse into a single point. This mathematically guarantees feature diversity without requiring stop-gradients or frozen networks.
3. The Method, Step-by-Step
LeWorldModel operates through a surprisingly straightforward pipeline (as visualized in Figure 1 of the paper):

- The Encoder: A small Vision Transformer (ViT) takes a raw pixel image and compresses it into a compact latent vector (e.g., 192 dimensions).
- The Predictor: Another Transformer takes this current latent vector, along with the agent’s action, and predicts the next latent vector.
- The Two-Term Loss: The entire system is trained end-to-end using only two signals:
- Prediction Loss (MSE): Make the predicted next state match the actual next encoded state.
- Anti-Collapse Loss (SIGReg): Project the batch of latent vectors onto random 1D lines and ensure their distribution looks like a bell curve (using the Epps-Pulley normality test).
- Latent Planning: At inference time, the model uses a Model Predictive Control (MPC) algorithm to simulate thousands of possible action sequences in this fast, compressed latent space, picking the one that gets the agent closest to its goal.

Key Experimental Results

The authors tested LeWM across 2D and 3D control tasks (Push-T, Reacher, TwoRoom, OGBench-Cube) against state-of-the-art JEPA models.
- Massive Speedup: Because LeWM encodes observations into ~200× fewer tokens than models relying on DINOv2, it achieved planning speeds up to 48× faster than DINO-WM, completing full planning steps in under one second.
- Competitive Task Performance: On complex manipulation tasks like Push-T, LeWM significantly outperformed PLDM (achieving an 18% higher success rate) and matched or beat DINO-WM, even when DINO-WM was given privileged robot-arm state data.
- Emergent Physics Engine: The authors used a “Violation-of-Expectation” (VoE) framework—a psychological test used on human infants. They showed the model a trajectory where an object suddenly teleported (violating physics). LeWM’s prediction error spiked dramatically, proving the latent space had genuinely learned the rules of physical continuity.
- Hyperparameter Simplicity: While PLDM requires tuning six sensitive loss weights, LeWM only requires tuning one (the weight of the SIGReg loss), making the training process smooth and monotonic.
A Critical Look: Strengths & Limitations
Strengths / Contributions
- Architectural Elegance: Reducing a complex stabilization problem down to a single statistical constraint (SIGReg) is a massive win for reproducibility. By ditching EMAs and stop-gradients, the optimization dynamics become transparent and mathematically sound.
- Democratizing World Models: At roughly 15 million parameters, LeWM can be trained from scratch on raw pixels in just a few hours on a single NVIDIA L40S GPU. This dramatically lowers the barrier to entry for world-model research.
- Interpretable Latents: The paper demonstrates “temporal latent path straightening”—an emergent phenomenon where latent trajectories naturally become smooth and straight over time. This proves the representations are capturing high-level dynamics rather than chaotic noise.
Limitations / Open Questions
- Struggles in Low-Complexity Environments: Ironically, LeWM performs slightly worse than baselines on the simplest environment tested (TwoRoom). Forcing a high-dimensional Gaussian distribution onto a dataset with very low intrinsic dimensionality (moving a dot in an empty room) forces the model to invent unnecessary structure, slightly degrading planning.
- Missing Fine-Grained Rotations: Probing experiments (Table 4) reveal that LeWM struggles to capture fine-grained rotational information (like block quaternions) from pixels alone. DINO-WM retains an edge here, likely because its foundation model encoder was pre-trained on 142 million diverse images.
- Reliance on Action Labels: Like most forward-dynamics world models, LeWM requires a dataset strictly annotated with actions. Moving toward action-free inverse dynamics will be necessary to leverage the massive amount of unlabelled video data on the internet.
Contribution Level: Significant Improvement. LeWM does not invent the JEPA paradigm, but it fundamentally cures its biggest headache. By proving that end-to-end, raw-pixel predictive architectures can be stabilized purely through statistical regularization—without massive compute or brittle heuristics—it sets a new, streamlined baseline for the entire field of latent-space planning.
Conclusion: Potential Impact
LeWorldModel is a breath of fresh air in a subfield that was becoming bogged down by complex architectural “tricks.” By ensuring that embeddings remain diverse via a simple Gaussian regularizer, the authors have shown that small, highly efficient world models can understand physics and plan complex tasks right out of the box.
This research will primarily benefit roboticists and reinforcement learning researchers who need lightweight, fast-planning world models that can be trained on domain-specific data without access to server farms. Looking ahead, if this stable JEPA architecture can be scaled up and pre-trained on wild, unannotated video datasets, it could serve as a powerful plug-and-play “common sense” physics engine for the next generation of autonomous agents.
- Title: [LeWM] How to Train Stable World Models from Pixels with Just Two Loss Terms
- Author: Jellyfish
- Created at : 2026-04-23 17:12:37
- Updated at : 2026-04-23 08:25:17
- Link: https://makepaperseasy.com/posts/20260423171237/
- License: This work is licensed under CC BY-NC-SA 4.0.
[SSF2020] Blurring to Compress Better: Google's Scale-Space Flow for Video Compression
Real-Time Video Rendering with 4D Gaussian Splatting
Smarter, Not Louder: How LLMs Cut Multi-Agent Communication by 53% While Boosting Performance
[Deformable 3D Gaussians] Bringing 3D Gaussian Splatting to Life for Real-Time Dynamic Scenes
[LISA] From 'Segment the Car' to 'Segment the Safest Place for a Toddler': LLMs Learn to Reason and See
[DrivingDojo] Why Can't Self-Driving AIs Turn Left? The Dataset for Smarter World Models
[SSF2020] Blurring to Compress Better: Google's Scale-Space Flow for Video Compression
Real-Time Video Rendering with 4D Gaussian Splatting
Smarter, Not Louder: How LLMs Cut Multi-Agent Communication by 53% While Boosting Performance
[Deformable 3D Gaussians] Bringing 3D Gaussian Splatting to Life for Real-Time Dynamic Scenes