Papers
Topics
Authors
Recent
2000 character limit reached

DPNet: Doppler Planning Network

Updated 6 December 2025
  • DPNet is a model-based learning framework that integrates Doppler LiDAR, neural Kalman filtering, and MPC for precise, real-time dynamic obstacle tracking.
  • The system employs D-KalmanNet, an adaptive neural Kalman filter that learns measurement gains to mitigate noise and accommodate rapid, non-constant-acceleration behavior.
  • DPNet shows significant performance gains, with up to 12 dB NMSE improvement over traditional filters on both highway and urban benchmarks.

Doppler Planning Network (DPNet) is a model-based learning framework designed for real-time motion planning in highly dynamic environments by integrating Doppler LiDAR data for both obstacle tracking and ego-motion planning. DPNet addresses limitations of prior methods in handling rapid obstacle motion by leveraging Doppler-derived velocity information, a custom neural Kalman filtering architecture (D-KalmanNet), and a Doppler-tuned model predictive control (DT-MPC) module. DPNet attains both high-frequency and high-accuracy behavior with minimal data requirements, as demonstrated on simulated and real-world datasets (Zuo et al., 29 Nov 2025).

1. Doppler LiDAR Integration for Dynamic Obstacle Tracking

At the core of DPNet is the use of Doppler LiDAR, which augments traditional ranging measurements with instantaneous pointwise velocity estimates. For each obstacle nn, DPNet constructs discrete-time, Gaussian state-space models (GSSMs) that encode planar position, velocity, and acceleration using a constant-acceleration motion model. The 6-dimensional state at time tt is:

xtn=[xtn cosθtnvtn cosθtnatn ytn sinθtnvtn sinθtnatn]R6\mathbf{x}_t^n = \begin{bmatrix} x_t^n \ \cos\theta_t^n\,v_t^n \ \cos\theta_t^n\,a_t^n \ y_t^n \ \sin\theta_t^n\,v_t^n \ \sin\theta_t^n\,a_t^n \end{bmatrix} \in \mathbb{R}^6

Here, (xtn,ytn)(x_t^n, y_t^n) is the central position, θtn\theta_t^n the heading, vtnv_t^n the speed, and atna_t^n the acceleration magnitude for obstacle nn. The transition matrix T\mathbf{T} is structured to enforce constant-acceleration kinematics with a sampling interval Δt\Delta t, and process noise wtnN(0,Q)w_t^n \sim \mathcal{N}(0, \mathbf{Q}).

The observation vector ytnR4\mathbf{y}_t^n \in \mathbb{R}^4 contains planar position and velocity, derived from Doppler-fused LiDAR point clusters. Each Doppler reading μti\|\bm{\mu}_t^i\| from point ii is projected into the tracking plane and aggregated to yield the obstacle’s velocity.

2. D-KalmanNet: Neural Kalman Filtering with Learned Gain

DPNet introduces D-KalmanNet, a neural Kalman filter variant that replaces the analytic Kalman gain computation with a learned, adaptive gain matrix. For each obstacle and timestep, the filter executes:

Prediction:

x^tt1=Tx^t1\hat{\mathbf{x}}_{t|t-1} = \mathbf{T} \hat{\mathbf{x}}_{t-1}

Measurement Update (learned gain):

Kt=RNN(x^tt1,y^tt1,yt,ht1)\mathbf{K}_t = \mathsf{RNN} \left( \hat{\mathbf{x}}_{t|t-1}, \hat{\mathbf{y}}_{t|t-1}, \mathbf{y}_t, \mathbf{h}_{t-1} \right)

x^t=x^tt1+Kt(yty^tt1)\hat{\mathbf{x}}_{t} = \hat{\mathbf{x}}_{t|t-1} + \mathbf{K}_t \left( \mathbf{y}_t - \hat{\mathbf{y}}_{t|t-1} \right)

The RNN is instantiated as a Gated Recurrent Unit (GRU) with a 64-dimensional hidden state; its input is the concatenation of the predicted state (6D), predicted observation (4D), and measurement (4D), for a total 14-dimensional input. The GRU output is processed by an MLP (64→32→24) with ReLU activations, producing the 6×46 \times 4 gain.

This learned gain enables robust compensation for model–data mismatch, rapid, non-constant-acceleration behavior, and partial observability due to real-world noise.

3. Doppler-Based Velocity Rectification and Measurement Fusion

DPNet incorporates a Doppler-based velocity rectification stage to address the high variability of pointwise radial velocities. For each obstacle, LiDAR returns are clustered, each Doppler reading is projected onto the planar velocity components using known geometry, and a noise-reducing averaging is performed:

(vx,tn,vy,tn)=1IniInμti˘cos(ιti)(v_{x,t}^n, v_{y,t}^n) = \frac{1}{|I_n|} \sum_{i \in I_n} \frac{\breve{\bm{\mu}_t^i}}{\cos(\iota_t^i)}

where InI_n indexes LiDAR points belonging to obstacle nn, and ιti\iota_t^i encodes the incidence angle. This fusion substantially reduces radial-velocity noise and facilitates more accurate downstream filtering.

The measurement model thus fuses position and planar velocity: ytn=Uxtn+vtn,vtnN(0,R)\mathbf{y}_t^n = \mathbf{U} \mathbf{x}_t^n + v_t^n , \quad v_t^n \sim \mathcal{N}(0, \mathbf{R}) with U\mathbf{U} projecting the 6-dimensional state into the observable 4 dimensions.

4. Model Predictive Planning: Doppler-Tuned MPC

Using the real-time obstacle motion tracks generated by D-KalmanNet, DPNet employs a Doppler-tuned model predictive control (DT-MPC) module. The predicted obstacle trajectories serve as input to the planner, which dynamically tunes MPC parameters at runtime. This allows online adaptation to environmental volatility, enhancing the agent’s reactivity to rapid, unpredictable obstacle motion. The lightweight, data-efficient architecture enables maintaining high planning frequency under computation and memory constraints.

5. Training Methodology and Empirical Results

D-KalmanNet is trained on AevaScenes real-world Doppler LiDAR sequences with ground-truth trajectories. For each training minibatch covering a window of HH future steps, the loss minimizes position (and optionally velocity) prediction errors:

L=nh=1Hx^t+htnxt+hn2+λhKt+hF2\mathcal{L} = \sum_{n} \sum_{h=1}^{H} \left\| \hat{x}^{n}_{t+h | t} - x^{n}_{t+h} \right\|^2 + \lambda \sum_{h} \| \mathbf{K}_{t+h} \|_F^2

Where λ\lambda is an L2L_2 regularizer on the gain. Adam is used for optimization (learning rate 10310^{-3}, weight decay 10510^{-5}), with training for 2000 epochs, batch size of 32, gradient clipping at norm 1.0, and early stopping based on validation MSE.

On the AevaScenes dataset (10 Hz, H=5H=5), D-KalmanNet achieves 35.80±8.62-35.80 \pm 8.62 dB NMSE on highway sequences and 45.00±7.35-45.00 \pm 7.35 dB NMSE in city, representing approximately 12 dB improvement over Doppler-aided Kalman filter baselines. The system maintains superiority at lower observation rates and longer prediction horizons. Per-step NMSE reveals especially large performance gaps for near-term forecasts (Step 1 NMSE: 48.8-48.8 dB for D-KalmanNet vs 16.5-16.5 dB for a baseline KF in city). Inference runs at 100 Hz for a single obstacle and above 15 Hz for 10 obstacles on Jetson Orin NX hardware (107 MB GPU memory).

6. Architectural and Algorithmic Innovations

DPNet advances standard neural Kalman filtering by tightly integrating Doppler-specific rectification, partial-observability handling, and a high-dimensional learned gain. Notable departures from prior filtering strategies include:

  • Doppler-based aggregation of noisy radial velocities into a single, less noisy planar velocity per obstacle, enabling robust and accurate measurement updates.
  • Explicit augmentation of the state-space to encompass heading uncertainty, velocity, and acceleration (rather than pure position-velocity).
  • RNN-based learning of the Kalman gain Kt\mathbf{K}_t, which adapts to rapid, non-constant-acceleration behaviors that lie outside model assumptions.
  • Lightweight design supporting end-to-end training and real-time inference on embedded platforms.

This combination enables DPNet to outperform fixed-model Kalman filtering and other learning-based baselines in both static and rapidly evolving environments (Zuo et al., 29 Nov 2025).

7. Benchmarking and Broader Impact

DPNet’s real-time tracking and planning capabilities are substantiated by extensive benchmarks on simulation and real-world data. Empirical superiority is established not just in low-level trajectory following but in system-level planning tasks involving rapid, unpredictable dynamic obstacles. The architecture demonstrates scalability to multiple obstacles with modest computational footprint and improved generalization over task-specific learning approaches.

A plausible implication is that DPNet, by harmonizing physically grounded motion models with learnable adaptation from high-framerate Doppler signals, forms a blueprint for next-generation autonomous agents navigating complex, dynamic scenes. The modular design is amenable to further extension, such as incorporating richer sensor modalities or more expressive control policies.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Whiteboard

Follow Topic

Get notified by email when new papers are published related to Doppler Planning Network (DPNet).