Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash 102 tok/s
Gemini 2.5 Pro 51 tok/s Pro
GPT-5 Medium 30 tok/s
GPT-5 High 27 tok/s Pro
GPT-4o 110 tok/s
GPT OSS 120B 475 tok/s Pro
Kimi K2 203 tok/s Pro
2000 character limit reached

Denoising World Model Learning

Updated 19 July 2025
  • Denoising World Model Learning (DWL) is a family of techniques that use multi-level denoising to construct world models resilient to noise, partial observability, and domain shifts.
  • It employs advanced methods like learnable wavelet transforms, spatio-temporal masking, and bisimulation losses to enhance feature invariance and generalization.
  • DWL approaches improve sim-to-real transfer and control tasks, offering practical benefits for robust reinforcement learning in dynamic, unpredictable settings.

Denoising World Model Learning (DWL) refers to a family of representation learning and reinforcement learning methodologies dedicated to constructing world models that are robust to noise, partial observability, distractors, and domain shift. By leveraging denoising techniques at the signal, representation, and policy-learning levels, DWL aims to produce latent state representations that are resilient to exogenous noise and encode the endogenous, task-relevant dynamics of the environment. This approach is motivated by the growing need for robust, generalizable artificial agents that function reliably under real-world conditions where sensory data and environmental factors are often unpredictable or substantially corrupted.

1. Core Concepts and Theoretical Foundations

DWL methodologies build on a range of denoising and representation learning principles:

  • Signal-level denoising uses classical and deep learning-based denoising approaches (e.g., learnable wavelet transforms) to preprocess raw sensory streams, removing unwanted noise before downstream modeling (Frusque et al., 2022).
  • Latent denoising involves learning world models whose internal representations are purified of irrelevant or exogenous information. This is achieved via approaches such as spatio-temporal masking, bisimulation-inspired losses, disentanglement regularization, and denoising auxiliary tasks (Poudel et al., 2023, Sun et al., 10 May 2024, Wang et al., 11 Mar 2025).
  • Denoising as auxiliary supervision incorporates explicit denoising or reconstruction losses to enforce invariance, for example using image denoising as an auxiliary task to aid contrastive representation learning in world models (Poudel et al., 2023).
  • Sim-to-real robustness is explicitly targeted in frameworks that use denoising to close the sim-to-real gap, especially in robotics, by ensuring the policy learns to act on latent representations robust to both simulated and real-world noise (Gu et al., 26 Aug 2024).

DWL unifies these strategies under a common purpose: training an agent to infer, plan, and act using high-fidelity representations that are minimally affected by noise and environmental variability.

2. Methodologies and Architectures

A variety of DWL architectures address different denoising challenges and target domains:

a) Signal-Level Denoising with Learnable Wavelet Packet Transforms

The Learnable Wavelet Packet Transform (L‑WPT) employs a tree-structured encoder-decoder kindred to classical WPT but enhanced with learnable parameters. Denoising is effected using a double sharp sigmoid activation that mimics soft thresholding in the wavelet domain, which can be adapted post-training to shifting noise profiles via a simple scaling modification to activation biases. This architecture achieves strong robustness across unseen noise intensities and signal classes, serving as a robust front-end for further world modeling (Frusque et al., 2022).

b) Model-Blind Temporal Denoising

DWL extends self-supervised video denoising via noise2noise learning, but addresses crucial pitfalls such as noise overfitting, occlusion, and lighting variation. Innovations like the twin sampler (which decouples input from target to prevent noise memorization), online denoising for better optical flow estimation, forward-backward consistency for accurate occlusion masking, and penalties based on measured lighting variation yield a state-of-the-art framework that generalizes to various noise types and architectures (Li et al., 2020).

c) Contrastive and Auxiliary Denoising in World Model Learning

ReCoRe proposes a world model that learns invariant features by combining contrastive InfoNCE losses—used to ensure the encodings are insensitive to style interventions—with auxiliary losses (such as denoising or depth reconstruction) that explicitly regularize invariance. The auxiliary denoising task forces the latent representation to recover the noise-free observation, providing a powerful regularizer against feature collapse and ensuring meaningful, robust representations for control (Poudel et al., 2023).

d) Spatio-Temporal Masking and Bisimulation

A Hybrid Recurrent State-Space Model (HRSSM) incorporates spatio-temporal masking (random cuboid masking of input sequences) and a bisimulation-inspired similarity loss that relates latent representation similarity to behavioral similarity. This is augmented with an explicit latent reconstruction loss between branches with and without masking. The hybrid design, combining a mask branch and an exponential moving average (EMA) raw branch, further stabilizes the jointly trained policy and world model (Sun et al., 10 May 2024).

e) Knowledge Transfer through Disentangled Latent Distillation

Disentangled World Models (DisWM) pretrain disentangled representations from distracting (noisy) videos using a β-VAE architecture. Semantic knowledge in the form of factorized latent variables is distilled into an action-conditioned world model via explicit KL divergence minimization. The process is complemented by further disentanglement regularization during online adaptation, resulting in effective cross-domain transfer and enhanced sample efficiency (Wang et al., 11 Mar 2025).

f) Denoising World Model Learning in Humanoid Locomotion

In complex robotics domains, DWL utilizes encoder–decoder structures with asymmetric actor-critic architectures to reconstruct full privileged state from noisy partial observations. Denoising losses, sparse latent representations, domain randomization, and active ankle control mechanisms collectively enable robust sim-to-real transfer and successful navigation of challenging terrains (Gu et al., 26 Aug 2024).

3. Training Protocols, Data, and Loss Functions

DWL systems typically employ the following procedural elements:

  • Self-supervised/noisy-to-noisy training: Ground-truth denoised data is unnecessary; instead, models are trained on noisy pairs using specially constructed loss functions, e.g., noise2noise L₂ loss (Li et al., 2020).
  • Twin sampling and masking strategies: Custom samplers and structured masking separate endogenous from exogenous or redundant content, essential for preventing overfitting to noise (Li et al., 2020, Sun et al., 10 May 2024).
  • Auxiliary and contrastive losses: Auxiliary tasks (denoising, depth, segmentation) are combined with contrastive learning objectives (InfoNCE) to avoid representation collapse and enforce feature invariance (Poudel et al., 2023).
  • Disentanglement and distillation losses: β-VAE losses and cross-domain KL distillation losses maintain interpretable and transferable representations (Wang et al., 11 Mar 2025).
  • Domain randomization and noise injection: Training regimes inject a variety of synthetic and environmental noises, as well as randomize physical parameters, to increase robustness, especially for sim-to-real scenarios (Gu et al., 26 Aug 2024).
  • Latent reconstruction and bisimulation loss: Ensures the alignment of representations and emphasizes behavioral equivalence (Sun et al., 10 May 2024).

4. Practical Applications and Empirical Results

DWL has shown significant impact across a broad range of real-world and simulated domains:

  • Audio and time-series processing: L‑WPT-based models outperform both wavelet shrinkage and other neural networks in background noise removal and generalize robustly to unseen audio conditions (Frusque et al., 2022).
  • Video denoising: Model-blind frameworks with twin sampling and online denoising consistently exceed baselines by 0.6–3.2 dB PSNR, achieving state-of-the-art results across noise types and datasets (Li et al., 2020).
  • Visual navigation and robotics: ReCoRe demonstrates improved out-of-distribution generalization and sim-to-real transfer, outperforming RAD, CURL, and DreamerV2 on visual navigation in iGibson (Poudel et al., 2023).
  • Reinforcement learning under distractors: HRSSM-based DWL achieves high average returns and low variance on DMC and ManiSkill tasks in the presence of complex, natural distractions (Sun et al., 10 May 2024).
  • Autonomous driving: DriveWorld achieves substantial gains in 3D object detection, mapping, tracking, and motion forecasting over previous pre-training paradigms, demonstrating the value of robust denoised 4D scene representations (Min et al., 7 May 2024).
  • Humanoid robotics: DWL enables robust zero-shot transfer of locomotion policies from simulation to a variety of complex real-world terrains (stairs, snow, uneven ground), achieving 100% success in extensive trials (Gu et al., 26 Aug 2024).
  • Sample efficiency and transfer: DisWM pretraining yields RL agents that rapidly adapt to domains with substantial underlying variation, demonstrated by superior performance across multiple transfer scenarios in DMC and MuJoCo tasks (Wang et al., 11 Mar 2025).

5. Ablations, Robustness, and Limitations

Ablation studies consistently demonstrate that DWL methods derive substantial benefit from each component:

  • Removing structured sampling or masking degrades performance by up to 3.7 dB PSNR (for video denoising) or substantially reduces sample efficiency and final return (for RL tasks) (Li et al., 2020, Sun et al., 10 May 2024, Wang et al., 11 Mar 2025).
  • Dedicated denoising auxiliary losses or depth reconstruction are found to be critical for preventing representation collapse—without them, both control and generalization degrade notably (Poudel et al., 2023).
  • Disentanglement regularization and latent distillation, when ablated, result in loss of semantic transfer and reduced adaptation in RL settings (Wang et al., 11 Mar 2025).

Identified limitations include:

  • Some methods, such as denoising as an auxiliary loss, may not be strictly style-invariant and can be less effective where geometric invariance is critical (Poudel et al., 2023).
  • Computational overhead varies: while some designs are lightweight (L‑WPT), others require recurrent networks, memory banks, or cross-attention, and performance at large scale or on hardware-constrained platforms may warrant further investment.
  • Joint training of representation, dynamics, and policy may introduce instabilities, partly addressed in the literature via hybrid and EMA-based architectures (Sun et al., 10 May 2024).

6. Outlook and Implications

Denoising World Model Learning continues to evolve as a critical component in deploying robust agents across real-world domains. Its foundational techniques—ranging from signal processing to probabilistic latent variable modeling, self-supervised denoising, disentanglement, and auxiliary regularization—intersect with the most pressing challenges in reinforcement learning and embodied AI. DWL frameworks are increasingly deployed for policy transfer across domains, sim-to-real learning in robotics, sample-efficient RL under distribution shift, and environments with heavy exogenous noise.

A plausible implication is that future world model architectures will intensify the integration of structured, interpretable denoising (e.g., adaptive masking, latent distillation) with large-scale self-supervision and domain adaptation. The continued success of these methods in practical applications—ranging from humanoid robotics and autonomous driving to video understanding and time-series analysis—suggests that denoising will remain a central pillar of robust world model learning research and its real-world deployment.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube