V-JEPA-2-AC: Video World Modeling for Robotics
- The paper demonstrates a two-stage model that decouples a frozen vision encoder from an action-conditioned predictor to enable zero-shot robotic manipulation.
- The method leverages a Vision Transformer pretrained on web-scale data and employs a compact Transformer for latent dynamics prediction using L₁ reconstruction loss.
- Empirical results show robust performance in grasping, reaching, and pick-and-place tasks, outperforming diffusion-based models in efficiency and control.
V-JEPA-2-AC is a two-stage, self-supervised video world modeling and planning system that leverages large-scale video representation learning and action-conditioned latent dynamics for zero-shot robotic manipulation. It is grounded in the V-JEPA 2 architecture and applies a frozen vision backbone for extracting latent video states, followed by post-training a compact action-conditioned predictor on robot data. The resulting world model enables image-goal planning in unseen physical environments using only monocular vision, without explicit task supervision or reward signals (Assran et al., 11 Jun 2025).
1. Model Architecture
V-JEPA-2-AC employs a bifurcated design:
- Stage 1: Encoder Pre-training
- Utilizes the action-free V-JEPA 2 encoder ; each RGB frame is transformed into a latent map .
- The encoder backbone is a Vision Transformer (ViT-g, up to $1$B parameters), pretrained on over $1$M hours of web-scale video and image data (VM22M), supporting up to $64$-frame clips and spatial resolution.
- Stage 2: Action-Conditioned Predictor
- Freezes and uses it to embed video frames during all subsequent learning and inference.
- Introduces a $300$M-parameter, $24$-layer Transformer () with block-causal attention, ingesting three token types per timestep:
- Patch features:
- Proprioceptive states: (Cartesian xyz, orientation, gripper)
- Actions:
- Input tokens are projected into a $1024$-dimensional hidden space and concatenated as chronological sequences with causal masking.
The predictor outputs the next latent patch map , matching the spatial dimensions of . There are no modifications to the encoder weights during post-training; the architecture is rigidly modular (Assran et al., 11 Jun 2025).
2. Training Objectives and Loss Functions
All learning proceeds without task labels or environmental rewards, relying exclusively on reconstruction losses:
- Teacher-Forcing ():
(Eq 2 in (Assran et al., 11 Jun 2025))
- Rollout Loss ():
(Eq 3)
- Total Loss:
(Eq 4)
There is no explicit contrastive or adversarial regularization; all supervision derives from the L₁ reconstruction of encoder-derived latents (Assran et al., 11 Jun 2025). A plausible implication is that modeling focuses solely on physical state prediction fidelity in a latent space, without explicit classification or discrimination tasks.
3. Latent World Model Formulation
- Latent State Extraction: For every video frame, produces a spatial map of latent embeddings corresponding to visual content.
- Action and State Conditioning: Actions are computed as differences in proprioceptive state vectors across timesteps, .
- Transition Modeling: The core latent dynamics are parameterized as:
- No additional regularizers modify the latent transition; only L₁ reconstruction loss applies (Assran et al., 11 Jun 2025).
This formalism enables an entirely end-to-end, action-conditioned latent space for video world modeling and planning. This suggests architectural efficiency, with all policy-relevant inference conducted in a shared, pretrained visual latent space.
4. Planning and Inference Mechanism
Image-goal planning proceeds in the latent space using an iterative trajectory optimization:
- Goal Encoding: The user specifies a goal image , embedded as .
- Current State Encoding: The latest frame and proprioceptive state are mapped to and .
- Trajectory Optimization: The method seeks a -step action sequence that minimizes
Solution:
This is optimized via the Cross-Entropy Method (CEM), which samples action sequences from Gaussian distributions, selects the highest-scoring trajectories, and refines the mean/variance over several iterations (typically 5–10).
- Execution: The robot performs only before replanning (receding-horizon / MPC) (Assran et al., 11 Jun 2025).
A plausible implication is that planning is robust to changes in environment and camera pose, as all inference is grounded in frozen visual latents.
5. Data, Training Protocols, and Deployment
- Pretraining Data: Mask-denoising JEPA on VM22M (web-scale video+image corpus, over $1$M hours); up to ViT-g scale, using $64$-frame, clips.
- Action-Conditioned Post-Training: Conducted with h unlabeled Droid robot video, $16$-frame clips @ resolution, no rewards or task labels, AdamW optimizer, $96$K total steps.
- Deployment: Franka Panda arms (RobotiQ gripper, monocular uncalibrated camera), zero-shot in two physical environments (no in-lab training), low-level operational-space control via planned -Cartesian actions (Assran et al., 11 Jun 2025).
This approach demonstrates domain generalization, leveraging tele-operation data from Droid for policy transfer without retraining.
6. Empirical Evaluation and Comparative Performance
Performance metrics and ablations establish V-JEPA-2-AC's capabilities:
| Method | Reach | Grasp (Cup/Box) | Reach w/ Obj (Cup/Box) | Pick-and-Place (Cup/Box) |
|---|---|---|---|---|
| Octo (BC) | 100% | 15% / 0% | 15% / 70% | 15% / 10% |
| V-JEPA 2-AC | 100% | 65% / 25% | 75% / 75% | 80% / 65% |
Meta-analysis with latent-diffusion world models (Cosmos):
- Cosmos: 80 samples, 10 refinements, 4 minutes/action (horizon=1), reach = 80%, manipulation = 0–20%
- V-JEPA 2-AC: 800 samples, 16 seconds/action, reach = 100%, manipulation = 60–80%
Ablations affirm the necessity of action-conditioning; unconditioned variants fail to infer control. Camera-pose dependence is linear in azimuth error, but can be addressed via linear calibration (Assran et al., 11 Jun 2025). A plausible implication is that latent-goal planning supports a broad spectrum of robot tasks without task-specific reward engineering or environment-specific retraining.
7. Context, Limitations, and Significance
V-JEPA-2-AC illustrates scalable video-LM world modeling with the following properties:
- Large-scale joint embedding from internet videos yields transferable latent representations for action-conditioned prediction.
- The separation of vision (frozen encoder) and dynamics (learned predictor) modularizes generalization and control.
- Zero-shot deployment in unseen labs, sans reward signals or task labels, achieves high success rates (65–80%) across grasp, reach-with-object, and pick-and-place tasks (Assran et al., 11 Jun 2025).
- Comparative study with diffusion-based world models evidences substantial improvements in efficiency and manipulation performance.
A plausible implication is that V-JEPA-2-AC's architecture and training protocol can generalize to more diverse environments and tasks by scaling video encoder pretraining and action-conditioned post-training, without revisiting reward-driven RL or in-situ data collection.
V-JEPA-2-AC, through its large-scale, decoupled training and latent planning formalism, represents a significant stage in the unification of self-supervised video representation learning with zero-shot robotic world modeling and control (Assran et al., 11 Jun 2025).