Papers
Topics
Authors
Recent
Search
2000 character limit reached

FILT3R: Latent State Adaptive Kalman Filter for Streaming 3D Reconstruction

Published 19 Mar 2026 in cs.CV, cs.AI, and cs.LG | (2603.18493v1)

Abstract: Streaming 3D reconstruction maintains a persistent latent state that is updated online from incoming frames, enabling constant-memory inference. A key failure mode is the state update rule: aggressive overwrites forget useful history, while conservative updates fail to track new evidence, and both behaviors become unstable beyond the training horizon. To address this challenge, we propose FILT3R, a training-free latent filtering layer that casts recurrent state updates as stochastic state estimation in token space. FILT3R maintains a per-token variance and computes a Kalman-style gain that adaptively balances memory retention against new observations. Process noise -- governing how much the latent state is expected to change between frames -- is estimated online from EMA-normalized temporal drift of candidate tokens. Using extensive experiments, we demonstrate that FILT3R yields an interpretable, plug-in update rule that generalizes common overwrite and gating policies as special cases. Specifically, we show that gains shrink in stable regimes as uncertainty contracts with accumulated evidence, and rise when genuine scene change increases process uncertainty, improving long-horizon stability for depth, pose, and 3D reconstruction, compared to the existing methods. Code will be released at https://github.com/jinotter3/FILT3R.

Authors (2)

Summary

  • The paper introduces an adaptive Kalman-style updating mechanism via FILT3R that ensures stable, long-horizon streaming 3D reconstruction.
  • It employs recursively propagated uncertainty by combining fixed measurement noise with adaptive process noise to balance memory retention against drift.
  • Experimental results on benchmarks demonstrate reduced error accumulation and lower computational overhead compared to prior methods.

FILT3R: Latent State Adaptive Kalman Filter for Streaming 3D Reconstruction

Problem Formulation: Streaming 3D Reconstruction and Updating Latent State

Streaming 3D reconstruction tasks require online maintenance of a persistent scene representation as video frames arrive, under constant memory and latency constraints. Contemporary recurrent architectures distill scene information into compact latent tokens updated by incoming frames. However, standard update rules—such as uniform overwrite (CUT3R) or heuristic gating (TTT3R)—fail to balance memory retention against adaptability, causing either catastrophic forgetting or excessive drift, particularly on video sequences exceeding the training horizon. Figure 1

Figure 1: FILT3R enables stable long-horizon streaming 3D reconstruction without resets. Top: input frames. Middle: backbone with different update policies. Bottom: FILT3R outperforms alternatives in maintaining stable geometry on long rollouts.

The core problem thus reduces to the design of a latent state update mechanism that adaptively trades off between trusting historical information versus incorporating new observations, ideally with robustness to distributional regime shifts beyond the trained data horizon.

FILT3R: Adaptive Latent Kalman-Style Token Filtering

FILT3R addresses this challenge by casting the latent token update as a stochastic state estimation problem using a per-token adaptive Kalman filter. The recurrent state is interpreted as a belief over scene latents, and the decoder’s candidate output as a noisy observation. FILT3R computes the Kalman-style gain via two uncertainty terms: a fixed measurement noise (reflecting decoder reliability) and an adaptively estimated process noise (reflective of scene dynamics), the latter inferred online from EMA-normalized temporal drift between consecutive candidate tokens. Figure 2

Figure 2: FILT3R overview. Latent memory and decoder candidate are fused using a per-token adaptive Kalman gain, which is driven by online noise estimation from drift and propagated variance.

Variance is propagated through time, so gains contract as the state grows more certain in stable regimes but increase sharply upon detecting transitions, leading to robust length generalization and principled balance between memory retention and adaptation.

Relationship to Existing Update Rules

FILT3R subsumes prior update rules as special cases: uniform overwrite corresponds to unit gain, fixed-gain gates to fixed Kalman gain, and heuristic gating to non-propagated, signal-based gating. In contrast to methods such as TTT3R or fixed EMA smoothing, FILT3R’s use of recursively propagated uncertainty yields an expanding memory horizon in stable conditions, with theoretical O(1/t)\mathcal{O}(1/t) gain decay [see Appendix]. Figure 3

Figure 3: Qualitative long-horizon streaming 3D reconstruction. FILT3R yields coherent geometry over very long rollouts; baselines degrade due to forgetting/drift.

Experimental Results

Long-Horizon 3D Reconstruction

On 7-Scenes and NRGBD benchmarks at sequence lengths up to 1000 frames—significantly beyond the 64-frame training horizon—FILT3R outperforms CUT3R and TTT3R by a wide margin in accuracy, completeness, and normal consistency. Notably, Point3R (which caches states) runs out of memory where FILT3R’s memory footprint remains constant.

Long-Horizon Pose and Depth

On TUM-RGBD, FILT3R reduces origin-aligned trajectory error (ATE_orig) by over 2× compared to TTT3R, specifically addressing progressive drift. On Bonn long video depth, FILT3R maintains significantly lower Abs Rel and higher δ<1.25\delta_{<1.25} compared to other streaming policies, confirming slower error accumulation as sequence length grows. Figure 4

Figure 4

Figure 4

Figure 4

Figure 4: Long-horizon metrics vs. sequence length. FILT3R’s errors grow more slowly than baseline methods in both pose and depth, consistently reducing cumulative drift.

Efficiency

Figure 5

Figure 5: Peak GPU memory vs. frames. FILT3R matches CUT3R’s constant memory profile, in contrast to full-attention and memory-caching approaches.

FILT3R imposes negligible computational and memory overhead compared to CUT3R, enabling efficient deployment over long horizons. Attention-based update gates (TTT3R) nearly double the memory footprint due to cache requirements.

Ablations and Mechanistic Insights

Ablations confirm: (i) adaptive process noise is necessary for rapid response to transitions; (ii) recursive variance propagation is critical for suppressing long-horizon drift; (iii) fixed-gain (EMA) gates can suppress drift only by globally over-smoothing, at the cost of responsiveness and local accuracy; (iv) adaptive measurement noise introduces instability due to correlated process and measurement estimates during transitions. Figure 6

Figure 6: Different update rules on the gain curve. FILT3R adapts its effective update gain based on process/measurement noise, unlike overwrite or fixed-gate approaches.

Mechanistically, FILT3R contracts updates during long stable periods (lowering gain floor) and selectively reopens when scene evidence indicates genuine change. Figure 7

Figure 7

Figure 7: FILT3R contracts updates in stable regimes and reopens at transitions. Effective gain and posterior variance dynamics highlight the adaptivity to temporal regime change.

Implications and Future Directions

FILT3R provides an interpretable, training-free, plug-and-play filter for latent token updates in recurrent 3D reconstruction. It facilitates significantly improved rollouts on video sequences of arbitrary length, demonstrating generalization well beyond training horizons. These outcomes indicate that principled uncertainty propagation is not only compatible with strong short-horizon metrics, but essential for stability in continuous deployment—a critical property for long-term embodied scene understanding and SLAM in real-world robotics.

Theoretically, FILT3R’s separation of process and measurement noise, and its explicit use of propagated uncertainty, offers a foundation for further advances: e.g., adaptive or learned process noise models, integration with parameter adaptation, and generalization to multi-modality or higher-order latent geometries. Future developments may combine FILT3R with differentiable filtering and end-to-end fine-tuning for task-specific optimality, or extend the adaptive filtering principle to other structured memory systems for continual RL or generative modeling.

Conclusion

FILT3R demonstrates that Kalman-style uncertainty-aware state fusion yields practical and significant improvements in the long-horizon stability and performance of streaming 3D reconstruction pipelines. By estimating process noise from internal temporal dynamics and justifying all adaptivity through process uncertainty—anchored on a fixed decoder reliability prior—FILT3R advances the state of the art in streaming scene reconstruction, with minimal computational cost and strong empirical validation across depth, pose, and reconstruction metrics.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We found no open problems mentioned in this paper.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 49 likes about this paper.