Multi-step EDMD for Nonlinear Systems
- Multi-step EDMD is a data-driven framework that leverages Koopman operator theory to approximate nonlinear dynamics via linear representations in a lifted function space.
- It minimizes long-horizon errors by directly learning multi-step state outputs, avoiding error compounding seen in recursive single-step methods.
- The framework employs advanced dictionary learning and sparsity techniques, yielding robust surrogates applicable to density forecasting, system identification, and model predictive control.
The multi-step Extended Dynamic Mode Decomposition (EDMD) framework is a family of data-driven numerical methods for forecasting, modeling, and controlling nonlinear dynamical systems via linear approximations in function space. By leveraging action of the Koopman operator on a finite set of observables, multi-step EDMD enables high-fidelity multi-horizon prediction and robust surrogate model construction for stochastic, deterministic, and controlled systems. Key advantages over classical single-step EDMD include direct minimization of long-horizon errors, avoidance of error compounding under recursive prediction, and improved tractability for real-time and control applications.
1. Mathematical Foundations and Koopman Lifting
The multi-step EDMD framework generalizes the spectral approach of Koopman operator theory to nonlinear systems, including both deterministic maps and stochastic differential equations (SDEs) with diffusion. Consider a continuous-time SDE
with initial distribution , drift , and noise , or a discrete-time system . The associated stochastic Koopman semigroup acts on observables by , defining a linear semigroup generated by the infinitesimal generator
with Fokker–Planck adjoint (Zhao et al., 2022). Under ergodicity, there exists a unique invariant density , and the Hilbert space inner product governs the spectral structure of and .
Deterministic systems use analogous constructions, with observables spanning a finite-dimensional subspace, and evolution in this lifted space governed by approximations of the infinite-dimensional Koopman operator (Schurig et al., 10 Nov 2025, Meda et al., 4 Apr 2025).
2. Finite-Dimensional EDMD Approximations and Multi-Step Forecasting
The core idea of EDMD is to approximate by a finite-dimensional matrix acting on the lifted state, using empirical data:
- Snapshot pairs are generated via dynamics and possible control inputs.
- The observables define the lift ; the data matrices are .
- Gram matrices are , .
- The Koopman approximation is , with eigenpairs giving approximate eigenvalues and eigenfunctions (Zhao et al., 2022, Meda et al., 4 Apr 2025).
Multi-step prediction proceeds by recursive iteration:
where the lifted state , selects physical states from the lifted coordinates, and recovers outputs or measurements (Meda et al., 4 Apr 2025). In stochastic density forecasting, spectral decompositions using generate multi-step probability evolution via or eigenmode propagation (Zhao et al., 2022).
3. Dictionary Design, Manifold Optimization, and Sparsity
The expressiveness and generalization of EDMD surrogates depend critically on the choice of observables. Standard bases include monomials, radial basis functions (RBFs), and neural networks (Meda et al., 4 Apr 2025). Dictionary learning is further refined by geometric optimization on the Grassmann manifold: selecting a subspace that minimizes multi-step forecast error via Riemannian optimization yields approximately invariant, low-dimensional subspaces that are robust to out-of-domain states (Schurig et al., 10 Nov 2025). The error metric integrates projection error over finite horizons and test initial conditions, and the optimization descends to Grassmannian geometry with Riemannian gradients and QR-based retraction.
Explicit dictionary pruning and structure discovery are enabled by including -type regularization in least-squares identification, which removes irrelevant observables and yields parsimonious, efficient surrogates (Wu et al., 17 Jan 2026). Parallel decomposition across states, steps, and dictionary rows supports scalable computation.
4. Multi-Step Least-Squares Identification and Error Control
A key advance of the multi-step EDMD framework is direct learning of the condensed multi-step state-output map. Rather than identifying an operator for single-step propagation and composing (which leads to error compounding and potential instability for ), the framework fits for each prediction step a linear map from dictionary lifts and past inputs to the -step-ahead state:
The identification is performed via convex least-squares:
with row-wise sparsity for dictionary pruning (Wu et al., 17 Jan 2026). This produces stable, direct multi-step surrogates whose error does not grow with prediction horizon . Theorem 3 in (Wu et al., 17 Jan 2026) provides, under Sobolev regularity and bounded-dictionary assumptions, high-probability error bounds independent of :
where is the dictionary size, the number of data, and the degree for polynomial lifts. In contrast, one-step EDMD bounds exhibit exponential-in- error blow-up when .
5. Applications, Numerical Performance, and Practical Considerations
Multi-step EDMD has been demonstrated for prediction, nonlinear system identification, density forecasting, and model predictive control (MPC) across a wide range of systems:
- Density forecast for SDEs: Approximates the semigroup generated by the Fokker–Planck operator, achieving weak convergence of the density representation as data and dictionary sizes increase (Zhao et al., 2022).
- Nonlinear system surrogates: For electric vehicle (EV) cabin climate, multi-step EDMD with RBF dictionaries achieves state RMSE and power RMSE for basis, outperforming polynomial and neural-network dictionaries at small (Meda et al., 4 Apr 2025).
- Koopman MPC: Direct multi-step identification solves the full condensed map for state and control over control horizon , enabling stable closed-loop control where one-step EDMD leads to diverging errors and controller instability. In benchmarks, multi-step EDMD achieves stable MSE under open and closed loop for , while one-step MSE diverges for due to spectral radius (Wu et al., 17 Jan 2026).
- Dictionary learning: Grassmannian-based shaping yields low-dimensional surrogates with out-of-domain generalization and large speedups in predictor evaluation (Schurig et al., 10 Nov 2025).
Per-step computational costs are governed by matrix-vector products and dictionary evaluation (with cost per basis). For –$40$, real-time execution on embedded platforms is practical (Meda et al., 4 Apr 2025).
6. Algorithmic and Implementation Summary
A generic multi-step EDMD workflow proceeds as:
- Dictionary selection: Choose functions —polynomial, RBF, or learned—appropriate to system structure and data availability.
- Data acquisition: Collect snapshot (state, next-state) or (trajectory) data under representative state and control distributions.
- Lifting and matrix assembly: Formulate , and associated Gram matrices; assemble multi-step input-output structures for supervised learning.
- Identification: Solve (regularized) least-squares for operator matrices, possibly under parallelizable decomposition across step and state ; perform dictionary pruning as necessary.
- Prediction: For given initial condition and input sequence, propagate via
up to desired horizon, reconstruct state and output via and matrices.
- Validation: Assess with trajectory RMSE, Consistency Index (CI), or finite-horizon forecast errors; compare performance with alternative dictionaries or identification schemes (Zhao et al., 2022, Meda et al., 4 Apr 2025, Schurig et al., 10 Nov 2025, Wu et al., 17 Jan 2026).
When relevant (e.g., Grassmannian dictionary learning), implement Riemannian optimization to shape active observables for improved multi-step invariance and forecast accuracy (Schurig et al., 10 Nov 2025). Closed-loop extensions with linear MPC/LQR or online adaptation using recursive least squares are direct, leveraging the lifted linear structure of the surrogate.
7. Convergence Guarantees and Limitations
Convergence properties are established in terms of increasing data (), basis set expressiveness (), and regularity of the lifting functions. For stochastic systems, EDMD approximations converge to the Galerkin projection of the Fokker–Planck semigroup in , and the truncated density forecast converges weakly to the true solution (Zhao et al., 2022). In deterministic and control settings, multi-step EDMD avoids error blow-up inherent to one-step recursion, with finite-time sample error scaling optimally in dictionary size and sample count (Wu et al., 17 Jan 2026). Overfitting and lack of generalization can arise with excessive dictionary dimension, motivating use of Grassmannian shaping and sparsity regularization (Schurig et al., 10 Nov 2025, Wu et al., 17 Jan 2026).
A plausible implication is that careful system-specific dictionary design, appropriate regularization, and direct multi-horizon identification are essential for the robust practical application of EDMD in high-dimensional settings and real-time control scenarios.