IDSplat: Self-Supervised 3D Dynamic Scene Reconstruction
- IDSplat is a self-supervised 3D Gaussian Splatting framework that reconstructs dynamic driving scenes by modeling static and dynamic objects as coherent instances using rigid transformations.
- It integrates zero-shot language-grounded video tracking with LiDAR anchoring and coordinated-turn smoothing to achieve physically consistent motion estimation and instance decomposition.
- The framework delivers competitive reconstruction quality and enables practical downstream applications such as object mask extraction and trajectory editing in autonomous driving scenarios.
IDSplat is a self-supervised 3D Gaussian Splatting framework designed for dynamic scene reconstruction with explicit instance-level decomposition and learnable motion trajectories, targeted for driving scenarios. It explicitly models dynamic objects as coherent instances undergoing rigid transformations through time, achieving efficient, annotation-free instance separation and scene rendering. The approach integrates zero-shot, language-grounded 2D video tracking with LiDAR anchoring in 3D and coordinate-turn smoothing for physically consistent object motion estimation. IDSplat achieves competitive reconstruction quality, robust generalization across diverse sequences and camera densities, and enables downstream manipulation such as object mask extraction and trajectory editing (Lindström et al., 24 Nov 2025).
1. Core Scene Representation
IDSplat represents a scene as a set of 3D Gaussian primitives, each associated with either the static background or a dynamic object instance. Each Gaussian, indexed by , is parameterized as follows:
- Occupancy ()
- World-frame 3D mean ()
- 3D covariance (, typically factored into scale and orientation)
- RGB base-color feature ()
- View-dependent feature ()
- Discrete instance label (; for background, for instances)
Each instance has a temporally-varying rigid transformation at frame , mapping canonical primitive centers to the world frame via
Rendering is performed via “splatting” following SplatAD’s differentiable point-based rasterizer. Each Gaussian projects to a 2D “splat” on the image plane; at pixel , the weight is
where is the camera projection and projected into image space. The per-pixel color,
is alpha-composited in front-to-back order; accumulated opacity is .
The same process admits an equivalent volumetric view:
with replaced by for each Gaussian.
2. Instance Decomposition Pipeline
Instance-level decomposition occurs in a multi-stage, self-supervised pipeline integrating vision and LiDAR data:
Zero-shot Masking and 3D Association
- Grounded-SAM-2, prompted with semantic labels (e.g., “car,” “truck”), generates 2D object masks per video frame.
- Masks are eroded (~few pixels) to reduce misalignments.
- Temporally nearest LiDAR points are projected into each image; each point is assigned a mask ID for candidate-instance association.
- DBSCAN clustering (minPts=10, 0.5~m) filters outliers, and the largest cluster per instance is retained.
Feature Correspondence and Rigid Pose Initialization
- For each instance and paired source/target frames :
- DINOv3 features are extracted per image and projected onto 3D points.
- Features are matched across frames by cosine similarity threshold .
- RANSAC+Umeyama estimates rigid transformation: sample 3 correspondences, estimate , count inliers (distance 0.1~m), accept largest-inlier hypothesis (ratio 0.5).
- On success, compose ; transformed points are merged back into the instance’s canonical point cloud set.
This staged approach yields temporally consistent 3D instance tracks from unannotated videos and LiDAR.
3. Motion Modeling and Temporal Smoothing
IDSplat models motion trajectories as per-frame rigid-body transforms for each instance . Primitive centers are transformed at each timestep as described above.
A factor-graph, implemented in GTSAM, imposes coordinated-turn (CT) smoothing:
- At every frame , the state vector is , with (forward speed) and (curvature).
- Measurement factors for each RANSAC pose :
with rotation std ~rad and translation std ~m.
- CT motion-model factors relate and :
The residual constrains to the predicted motion from .
- Priors include random walks for , small roll/pitch priors (~rad), and moderate curvature ().
- Outlier rejection: measurement factors with large whitened error () are dropped iteratively (up to 10 L–M solves).
This scheme mitigates pose misalignment and tracking errors, yielding physically plausible, temporally smooth trajectories.
4. Loss Functions and Joint Optimization
IDSplat employs joint optimization of all representation and motion parameters, driven by compositional loss terms:
with
- , where and are losses on rendered range/intensity, penalizes opacity before true range, is cross-entropy for LiDAR “ray drops.”
- , regularizing occupancy and covariance scale.
Optimization variables encompass all Gaussian parameters, decoder weights, sensor embeddings, and dynamic instance poses . Gaussians and decoders are optimized using Adam for 30k steps, with multi-scale image sampling and scheduled learning rates. Instance poses are initialized with CT-smoothing and further refined via the full loss gradients, integrating both photometric and LiDAR signals.
5. Experimental Results and Evaluation
IDSplat achieves state-of-the-art performance in self-supervised, instance-decomposed scene reconstruction on major autonomous driving benchmarks:
Novel View Synthesis (Waymo Open Dataset)
Comparison metrics across different settings:
| Method | PSNR | SSIM | LPIPS | DPSNR |
|---|---|---|---|---|
| IDSplat (DeSiRe-GS) | 30.61 | 0.897 | 0.163 | 28.49 |
| DeSiRe-GS (self-sup.) | 28.76 | 0.873 | 0.193 | 26.26 |
| Supervised SplatAD | 30.80 | 0.900 | 0.160 | 28.97 |
In the challenging AD-GS protocol (front camera), IDSplat matches or slightly lags supervised SplatAD while outperforming self-supervised AD-GS.
LiDAR Reconstruction
- On dynamic Waymo scenes, IDSplat attains depth RMSE ~m, intensity error $0.056$ (SplatAD: $0.055$), ray-drop accuracy ( for SplatAD), and Chamfer distance $1.16$ ($0.98$ for SplatAD), indicating close correspondence to supervised performance.
Generalization and Robustness
IDSplat remains robust under sparse view regimes, maintaining DPSNR ~dB at only 25% input views, where other self-supervised methods degrade by ~dB. On PandaSet (6 cameras, 50% train, annotation-free), IDSplat nearly matches the best supervised baseline in all metrics.
Qualitative Characteristics
IDSplat produces stable, per-instance masks suitable for object editing. It supports downstream applications such as object removal and trajectory rearrangement at test-time, demonstrated by stable masking and editing results.
6. Context, Significance, and Implications
IDSplat advances instance-aware dynamic scene reconstruction without reliance on object trajectory or mask annotations, aligning with the growing need for scalable, sensor-realistic simulation in autonomous driving. Its explicit instance modeling via rigid transformations improves separation of static and dynamic components relative to undifferentiated or purely time-varying representations. The integration of physically consistent coordinated-turn smoothing addresses typical pose drift and tracking errors, enhancing motion realism for simulation or editing.
A plausible implication is the practical deployment of IDSplat in large-scale, self-supervised pipelines, where annotation costs and re-training must be minimized. Its robustness to changing view densities and capacity to generalize to new driving environments further broaden its applicability. The instance-control afforded by the explicit decomposition offers new avenues in interactive simulation, dynamic object behavior research, and virtual scene manipulation.
For implementation and further technical details, refer to the IDSplat primary source (Lindström et al., 24 Nov 2025).