Papers
Topics
Authors
Recent
2000 character limit reached

Goal-Conditioned World Model (GCWM)

Updated 3 January 2026
  • Goal-Conditioned World Model is a reinforcement learning framework that conditions state transitions on explicit goal inputs for flexible, long-horizon planning.
  • It leverages deep, stochastic recurrent state-space models with dedicated encoders, transition modules, and decoders to handle high-dimensional data like images.
  • Empirical results demonstrate GCWM’s success in tasks such as robotic manipulation and navigation, using both forward and backward modeling strategies to improve goal-reaching.

A Goal-Conditioned World Model (GCWM) is a class of dynamical models in reinforcement learning (RL) and control that explicitly parametrize future or counterfactual predictions with respect to a specified goal state. Unlike standard world models, which forecast transitions solely based on the current state and action, GCWMs model trajectories, transitions, or value estimates conditioned on the achievement of arbitrary target or goal configurations. This enables agents to plan and execute actions towards novel goals specified at inference time, supporting flexible goal-reaching, manipulation, navigation, and compositional reasoning in high-dimensional, often visual, settings.

1. Mathematical Foundations and Model Families

GCWMs generalize classical state-transition models by introducing goal variables as explicit inputs to the transition or prediction functions. Let stSs_t \in \mathcal{S} be the state, atAa_t \in \mathcal{A} the action, and gGg \in \mathcal{G} the specified goal. The canonical GCWM parametrizes a conditional transition distribution pϕ(st+1st,at,g)p_\phi(s_{t+1} \mid s_t, a_t, g), or, when using latent space representations, pϕ(zt+1zt,at,g)p_\phi(z_{t+1} \mid z_t, a_t, g) for encoding z=Φ(s)z = \Phi(s). In practice, models are often implemented as stochastic recurrent state-space models (RSSMs) with encoder (for sts_t), goal encoder, goal-conditioned dynamics, and reconstructive decoders (Duan et al., 2024, Zhou et al., 29 Dec 2025).

Distinct GCWM variants include:

  • Forward GCWM: Models pϕ(zt+1zt,at,g)p_\phi(z_{t+1} \mid z_t, a_t, g), as in (Duan et al., 2024).
  • Backward GCWM: Models the distribution over predecessor latents pθ(zt1zt,at1)p_\theta(z_{t-1} \mid z_t, a_{t-1}) to generate backward traces, which supports the construction of goal-reaching datasets without external rewards (Höftmann et al., 2023).
  • Hybrid or Value-Oriented GCWM: Models value functions or feasibility indicators V(s;g)V(s;g) or Q(s,a;g)Q(s,a;g) jointly with latent representations, supporting implicit prediction of reachability or progress to a goal (Bagatella et al., 2023, Nasiriany et al., 2019).

This explicit goal-conditioning enables planning and policy learning toward arbitrary, possibly unseen, goals at test time.

2. Model Architectures and Training Objectives

GCWMs are instantiated with deep neural architectures tailored to high-dimensional state and goal spaces, often including visual observations:

Generic GCWM training minimizes a combination of the following (with coefficients α,β\alpha, \beta): L=Lrecon+αLtrans+βLKLL = L_{\text{recon}} + \alpha L_{\text{trans}} + \beta L_{\text{KL}} where:

  • Lrecon=E[logpθ(xtht,zt)]L_{\text{recon}} = -\mathbb{E}[\log p_\theta(x_t|h_t, z_t)] (reconstruction loss)
  • Ltrans=E[zt+1z^t+12]L_{\text{trans}} = \mathbb{E}[\,\|z_{t+1} - \hat z_{t+1}\|^2\,] (transition loss)
  • LKL=E[DKL(q(ztht,et)pϕ(ztht1,at1,g^))]L_{\text{KL}} = \mathbb{E}[\,D_{KL}(q(z_t|h_t,e_t)\|p_\phi(z_t|h_{t-1}, a_{t-1}, \hat g))\,] (regularization) as in (Duan et al., 2024). Specialized architectures like backward world models train using KL-divergence between predicted and encoded predecessor latents (Höftmann et al., 2023).

Novel variants, such as Act2Goal's joint flow-matching for vision and actions, minimize vector-field losses for continuous-time latent transitions (Zhou et al., 29 Dec 2025).

3. Exploration, Subgoal Selection, and Data Augmentation

The performance and generalization of GCWMs are critically dependent on the diversity and coverage of the trajectories used during training. Major approaches to address this include:

  • Unconstrained Goal Navigation (MUN): Periodically discovers maximally distinct subgoals via Farthest Point Sampling (FPS) on actions, and fills replay buffers with bidirectional transitions between these “subgoal” states. This broadens the data distribution and mitigates model misgeneralization between states seen only in unidirectional sequences (Duan et al., 2024). The MUN loop alternates between subgoal mining (Distinct Action Discovery), bidirectional navigation, and model/policy updates.
  • Backward Trajectory Generation: Simulates reversible predecessor rollouts from the goal using a backward GCWM, then organizes the resulting transitions into graphs for shortest-path improvement and dataset distillation for policy learning (Höftmann et al., 2023).
  • Curiosity-Driven Exploration: Maximizes model predictive uncertainty (curiosity) using an ensemble world model, resulting in exploratory trajectories that cover the state space and serve as substrate for offline goal-conditioned planning (Bagatella et al., 2023).

These strategies directly address the challenge that GCWMs often fail to generalize transitions across sub-trajectories not observed in on-policy data.

4. Planning, Policy Learning, and Deployment

GCWMs support model-based planning and policy optimization in latent or observed spaces, conditioned on arbitrary goals:

  • Latent Imagination: With a trained GCWM, agents “imagine” trajectories from the current latent ztz_t to the goal zgz_g by rolling out the world model under sampled or optimized action sequences (Duan et al., 2024, Zhou et al., 29 Dec 2025).
  • Subgoal Planning: Composite methods (e.g., LEAP in (Nasiriany et al., 2019)) plan via a chain of optimally placed latent subgoals z1:Kz_{1:K}, where each is constrained to lie in the high-density region of the immersive latent prior (preventing unrealistic transitions). Cross-Entropy Method (CEM) is used for optimization in latent space.
  • Graph-Based Value Aggregation: Offline planning can address value estimation artifacts (e.g., local optima) by constructing graphs connecting sampled visited states and aggregating short-horizon value estimates to correct both local and global estimation errors (Bagatella et al., 2023).
  • Imitation Learning: Backward GCWM-generated trajectories are pruned via shortest-path algorithms, and behavior cloning is performed on edges that strictly decrease path length to the goal, ensuring loop-free, efficient training data (Höftmann et al., 2023).
  • Multi-Scale Temporal Control: In Act2Goal, generated trajectories are temporally decomposed into proximal and distal anchors using Multi-Scale Temporal Hashing, supporting fine-grained closed-loop feedback and long-horizon task consistency (Zhou et al., 29 Dec 2025).
  • Policy Optimization: Policy gradients or actor-critic methods are applied on imagined trajectories, sometimes using learned temporal-distance rewards within latent space (Duan et al., 2024).

A summary of distinct GCWM-enabled planning and policy learning frameworks is presented below:

Method/Paper Planning Modality Policy Learning
GCWM+MUN (Duan et al., 2024) Latent rollouts over model Model-based actor-critic
Backward GCWM (Höftmann et al., 2023) Backward graph, shortest path Behavior cloning
LEAP (Nasiriany et al., 2019) Latent subgoal planning Goal-conditioned TD3
GCWM+Aggregation (Bagatella et al., 2023) Graph-augmented MPC Offline value training
Act2Goal (Zhou et al., 29 Dec 2025) Visual diffusion/flow Cross-attention policy
WMNav (Nie et al., 4 Mar 2025) VLM-based subtask decomposition VLM-prompted reasoning

5. Applications and Empirical Results

GCWMs have been validated across a range of challenging domains:

  • Robotic Manipulation: Act2Goal demonstrates strong zero-shot generalization in real-robot block stacking and manipulation, improving success rates from 30% to 90% on out-of-distribution tasks within minutes of self-supervised finetuning (Zhou et al., 29 Dec 2025).
  • Navigation: GCWM+MUN achieves ≈95% success in unconstrained navigation between any permutation of 15 subgoals in 3-Block Stacking, versus <60% for baselines. High generalization holds in Ant-Maze and Walker with hundreds of start-goal pairs (Duan et al., 2024).
  • Vision-Based Maze Solving: Backward GCWM consistently enables agents to reach single or multiple goals from all start states in high-dimensional maze settings, achieving up to 95% success (Höftmann et al., 2023).
  • Object-Goal Navigation (Embodied AI): WMNav integrates Vision-LLMs into a zero-shot GCWM that deploys a curiosity-based value map, yielding +3.2% SR improvements on HM3D and +13.5% SR on MP3D compared to prior approaches (Nie et al., 4 Mar 2025).
  • Offline Planning from Curiosity: Combining model-based planning with value aggregation yields up to 70% zero-shot success on long-horizon offline navigation, stratified by local/global value artifact correction (Bagatella et al., 2023).

A summary table of specific GCWM domains and benchmark results:

Paper Domain (Env) Success/Key Finding
(Duan et al., 2024) Block Stacking, Mazes ~95% subgoal navigation, SOTA GCRL
(Zhou et al., 29 Dec 2025) Manipulation (real robot) 30→90% zero-shot OOD success
(Höftmann et al., 2023) Visual Mazes Up to 95% all-starts-to-goal
(Bagatella et al., 2023) Maze_Large, Pinpad 70% (graph aggregate), SOTA zero-shot
(Nie et al., 4 Mar 2025) ObjectNav (HM3D, MP3D) +3.2–13.5% SR/SPL gain (zero-shot)

6. Limitations, Open Questions, and Future Directions

Empirical and theoretical analyses have established persistent and open challenges for GCWM development:

  • Generalization: GCWMs' predictive accuracy deteriorates for state transitions not densely covered during training, especially for backward or cross-trajectory transitions (Duan et al., 2024). Bidirectional/broad exploration (e.g., MUN, curiosity) is required for high generalization.
  • Subgoal Discovery: Methods such as Distinct Action Discovery (DAD) may select infeasible or irrelevant subgoals in high-dimensional or constrained domains. Robust automated discovery of meaningful subgoals remains unresolved (Duan et al., 2024).
  • Value Estimation Artifacts: Local and global optima in learned value landscapes can impede planning. Aggregate graph-based smoothing and sufficiently long-horizon planning are effective mitigations, but not closed-form guarantees exist (Bagatella et al., 2023).
  • Actionability of World Model Outputs: Many models require explicit planning at inference and may not fully exploit sample efficiency gains for direct goal-directed policy learning.
  • Vision-Language Integration: In systems such as WMNav, all world-modeling relies on frozen VLM inference, with no end-to-end gradient-based improvement. While this accelerates deployment and leverages large-scale knowledge, it constrains possible adaptation (Nie et al., 4 Mar 2025).

A plausible implication is that advances in scalable bidirectional data augmentation, uncertainty-aware planning, and hierarchical compositional modeling will remain central to future GCWM research. Cross-fertilization with model-free algorithms and dynamic subgoal selection also represent active areas.

GCWM methodology sits at the interface of model-based RL, goal-conditioned policy learning (GCRL), and hierarchical planning:

  • Latent Subgoal Planning: Similar in principle to hierarchical RL, but with latent variable models constraining subgoal feasibility and transitions (VAE-based (Nasiriany et al., 2019)).
  • Model-Based Imagination: Leveraging RSSMs and Dreamer-like actor-critic frameworks for planning in latent space (Duan et al., 2024).
  • Offline and Reward-Free Learning: Several GCWM techniques (backward modeling, curiosity, hindsight relabeling) enable policy or value extraction in the absence of shaped rewards (Höftmann et al., 2023, Bagatella et al., 2023, Zhou et al., 29 Dec 2025).
  • Vision-Language and Modular World Models: WMNav demonstrates the integration of pre-trained VLMs into online, goal-conditioned reasoning by decoupling perception, memory, planning, and action modules (Nie et al., 4 Mar 2025).

GCWMs offer a modular and expressive formalism for formulating, planning, and learning in diverse long-horizon, goal-driven tasks across simulated and real domains, with ongoing research addressing model generalization, planning efficiency, and interactive autonomy.

Whiteboard

Topic to Video (Beta)

Follow Topic

Get notified by email when new papers are published related to Goal-Conditioned World Model (GCWM).