Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 64 tok/s
Gemini 2.5 Pro 47 tok/s Pro
GPT-5 Medium 27 tok/s Pro
GPT-5 High 31 tok/s Pro
GPT-4o 102 tok/s Pro
Kimi K2 206 tok/s Pro
GPT OSS 120B 463 tok/s Pro
Claude Sonnet 4.5 35 tok/s Pro
2000 character limit reached

Latent Action Policy Learning (LAPO)

Updated 4 October 2025
  • Latent Action Policy Learning is a framework that infers unobserved action representations from state transitions to enhance reinforcement and imitation learning.
  • It decouples policy optimization from action modeling by employing auxiliary latent spaces via techniques like VAEs, diffusion models, and contrastive encoders.
  • The approach improves sample efficiency and transferability across tasks by leveraging unlabeled behavioral data and minimal supervision to align latent and real actions.

Latent Action Policy Learning (LAPO) is a research framework and algorithmic paradigm exploring the use of latent—i.e., unobserved and inferred—action representations within sequential decision-making models, particularly in reinforcement learning and imitation learning. The central objective is to decouple policy optimization and action modeling by introducing an auxiliary, compact, and often discrete latent action space. This latent space is either inferred directly from observational data (such as raw video or state trajectories) or engineered to simplify, unify, or regularize policy search. Recent research demonstrates that LAPO enables highly sample-efficient learning, robustness to partial supervision, and enhanced transferability across tasks, agents, or embodiments, particularly when access to explicit action labels is limited or costly.

1. Core Principles and Motivating Paradigms

LAPO is underpinned by the hypothesis that behaviorally meaningful actions can be learned as latent causes of observed state transitions, without requiring direct annotation of action labels. This is formalized in algorithms that use pairs of observations, or sequences thereof, to train an inverse dynamics model (IDM) that infers a latent action ztz_t capturing the essential causal factor explaining the transition (ot,ot+1)(o_t, o_{t+1}) (Schmidt et al., 2023). Complementary forward dynamics models (FDMs) predict future states when provided with inferred latent actions, ensuring these actions remain consistent with system dynamics.

A variety of generative modeling techniques are employed to parameterize the latent action space:

By treating action labels as latent variables, LAPO frameworks leverage vast, unlabeled behavioral datasets—such as video of expert agents, robots, or humans—and can subsequently map latent actions to ground-truth action spaces using a minor amount of supervised data or minimal environment interactions.

2. Methodological Taxonomy

Latent Action Mining and Policy Learning

Initial LAPO approaches first mine latent actions by modeling the conditional distribution of next state given current state and hypothesized latent action:

Lmin=minzΔtG(Ep(st),z)2\mathcal{L}_{\text{min}} = \min_{z} \| \Delta_t - G(E_p(s_t), z) \|^2

where GG is a generative forward dynamics model parameterized by a state embedding EpE_p (Edwards et al., 2018). Simultaneously, a policy πω(zst)\pi_\omega(z|s_t) estimates the likelihood of each latent action given the current state. Policy learning is then framed as maximizing concordance between predicted next states (under marginalized latent action distributions) and expert transitions:

Lexp=st+1s^t+12s^t+1=zπω(zst)G(Ep(st),z)\mathcal{L}_{\exp} = \| s_{t+1} - \hat{s}_{t+1} \|^2 \qquad \hat{s}_{t+1} = \sum_z \pi_\omega(z|s_t) G(E_p(s_t), z)

Action Remapping and Decoding

Since latent actions have no inherent grounding in the real action space, a remapping procedure aligns them to actual control commands:

zt=argminzst+1G(Ep(st),z)2z_t = \arg\min_z \| s_{t+1} - G(E_p(s_t), z) \|_2

A classifier πξ(az,Ea(st))\pi_\xi(a|z, E_a(s_t)) is then trained, on limited labeled data, to decode latent actions into true actions via cross-entropy supervision. This two-stage paradigm enables robust behavior imitation from observations alone, with limited real-world feedback (Edwards et al., 2018, Schmidt et al., 2023).

Latent Action Space Construction

Modern methods extend this pipeline through:

  • The use of VAEs/CVAEs to model the conditional distribution p(as,z)p(a|s, z), where sampling zz ensures coverage of the support seen in the dataset, alleviating extrapolation and out-of-distribution errors in offline RL (Zhou et al., 2020).
  • Diffusion-based models capturing latent actions as distributions over temporally extended trajectories or multi-step behaviors (Li, 2023, Tan et al., 12 Mar 2024).
  • Contrastive or alignment-based embedding strategies, unifying the latent action spaces across different embodiments (human/robot/parallel grippers) (Bauer et al., 17 Jun 2025).

Temporal and Structural Abstraction

LAPO methods often factor the RL or imitation learning pipeline through latent variables at the level of individual actions (per-step), skills (multi-step chunk), or even generating full trajectory segments in latent space. This enables temporal abstraction and shorter credit assignment in RL (Li, 2023, Tan et al., 12 Mar 2024).

3. Theoretical Properties and Identifiability

A rigorous theoretical foundation for LAPO has been established (Lachapelle, 1 Oct 2025), specifying desiderata for the learned latent action representation:

  1. Determinism: For every observed state/action pair, the IDM assigns a unique latent action (Dirac-delta property).
  2. Disentanglement: The latent action depends only on the underlying ground-truth action, not spurious details of the observation.
  3. Informativeness (Injectivity): The mapping from ground-truth actions to latent codes is injective; each real action has a distinct latent counterpart.

Under entropy-regularized objectives,

minf,q Ep(x,x)[Ea^q(x,x)xf(x,a^)22+βH(q(x,x))]\min_{f, q} \ \mathbb{E}_{p(x,x')} \left[ \mathbb{E}_{\hat{a}\sim q(\cdot|x,x')} \| x' - f(x, \hat{a}) \|_2^2 + \beta H(q(\cdot|x,x')) \right]

the LAPO framework provably identifies latent action representations meeting these criteria in the limit of sufficient data and under continuity/topology assumptions. Notably, discrete latent spaces are found to satisfy these desiderata efficiently in practice.

The statistical benefit is the ability to transfer policies learned in the latent space to real action spaces using minimal ground-truth action data—a property of critical importance for scaling to domains where annotated actions are prohibitively costly.

4. Empirical Performance and Applications

Extensive experiments in both classic control and high-dimensional domains validate LAPO’s practicality:

  • In low-dimensional settings (Cartpole, Acrobot, Mountain Car), LAPO achieves near-expert performance in fewer than 100 environment interactions (Edwards et al., 2018).
  • In visual domains (CoinRun, D4RL-V, LIBERO), latent-action-based approaches outperform prior baselines when trained on image data, including robustness to partial action supervision (Edwards et al., 2018, Tharwat et al., 22 Sep 2025).
  • For dialog systems (Zhao et al., 2019, Lubis et al., 2020), using latent actions reduces the RL horizon, improves dialog quality, and prevents catastrophic language degradation seen in word-level RL.
  • In multi-embodiment and cross-robot control, a unified latent space supports improved generalization and higher success rates, even in the presence of significant morphology gaps (Bauer et al., 17 Jun 2025, Zheng et al., 22 Mar 2025).
  • LAPO methods grounded in world modeling via Recurrent State-Space Models enable training directly from video, supporting sample-efficient transfer to real robotic manipulation tasks (Tharwat et al., 22 Sep 2025).

Latent action learning has also enabled advances in reasoning efficiency for LLMs (via length-adaptive policy optimization), allowing models to internalize resource allocation strategies and produce concise yet accurate outputs across problem complexities (Wu et al., 21 Jul 2025).

5. Challenges, Extensions, and Open Problems

LAPO faces significant challenges in realistic settings with distractors—environment changes not causally related to agent actions:

  • Under distractor-correlated noise (e.g., dynamic backgrounds), naïve LAPO may encode spurious transitions as latent actions, degrading downstream learning (Nikulin et al., 1 Feb 2025).
  • Techniques such as multi-step prediction, abandoning quantization, increasing latent dimensionality, and incorporating even minimal action supervision (as little as 2.5% labeled data) can mitigate this, restoring or improving latent action quality by factors up to 8×8\times relative to unsupervised baselines (Nikulin et al., 1 Feb 2025).

Object-centric approaches, which decompose observations into semantically meaningful object slots (via self-supervised video slot attention), further enhance latent action robustness. By focusing the inverse/forward dynamics modeling on task-relevant objects, proxy-action recovery and policy imitation become less sensitive to distractors, resulting in a 2.7×2.7\times improvement in latent action quality and up to 2.6×2.6\times higher imitation returns in visually rich environments (Klepach et al., 13 Feb 2025).

Identifiability remains a topic of active investigation, with entropy regularization and topological assumptions identified as key for guaranteeing proper mapping between latent and real action spaces (Lachapelle, 1 Oct 2025). Future LAPO research is directed towards:

  • Scaling to web-scale, unlabeled video corpora.
  • Developing more robust objective functions for distractor robustness.
  • Integrating structured, object-centric representations.
  • Extending transferability to a broader class of multi-task, multi-embodiment, and few-shot settings.
  • Deepening the theoretical guarantees regarding identifiability and modularity.

6. Codebases, Benchmarks, and Practical Implementation

Open-source code for seminal LAPO methods is made available, supporting reproducibility and further research. The ILPO method’s source (Edwards et al., 2018) provides reference implementations for both small-scale (classic control) and visual domains, with architectures amenable to adaptation for continuous control or higher-dimensional embedding schemes.

LAPO is actively evaluated on standard RL and imitation learning benchmarks, including D4RL, V-D4RL, Distracting Control Suite (DCS), Meta-World, CoinRun, Robomimic, MultiWoz, LIBERO, and others. Evaluation metrics include normalized episode return, task success rate, latent action probe accuracy (via linear probing to recover ground-truth actions), and cross-embodiment skill transfer rates.

Empirically, LAPO frameworks demonstrate substantial improvements over behavior cloning from observation, word-level reinforcement learning, and standard behavior-constrained offline RL—especially in settings with limited or partially observed action label data, or in environments requiring efficient policy adaptation and transfer.


Latent Action Policy Learning represents an overview of generative modeling, causality-aware dynamics modeling, and sample-efficient policy optimization, enabling data-driven learning of control behaviors from high-dimensional, partially labeled, or unlabeled datasets. Recent advances have established both strong empirical results and an emerging theoretical framework, with ongoing research addressing robustness, scalability, and identifiability in increasingly complex embodied and multi-modal real-world settings.

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Latent Action Policy Learning (LAPO).