Papers
Topics
Authors
Recent
Search
2000 character limit reached

Multi-step EDMD for Nonlinear Systems

Updated 24 January 2026
  • Multi-step EDMD is a data-driven framework that leverages Koopman operator theory to approximate nonlinear dynamics via linear representations in a lifted function space.
  • It minimizes long-horizon errors by directly learning multi-step state outputs, avoiding error compounding seen in recursive single-step methods.
  • The framework employs advanced dictionary learning and sparsity techniques, yielding robust surrogates applicable to density forecasting, system identification, and model predictive control.

The multi-step Extended Dynamic Mode Decomposition (EDMD) framework is a family of data-driven numerical methods for forecasting, modeling, and controlling nonlinear dynamical systems via linear approximations in function space. By leveraging action of the Koopman operator on a finite set of observables, multi-step EDMD enables high-fidelity multi-horizon prediction and robust surrogate model construction for stochastic, deterministic, and controlled systems. Key advantages over classical single-step EDMD include direct minimization of long-horizon errors, avoidance of error compounding under recursive prediction, and improved tractability for real-time and control applications.

1. Mathematical Foundations and Koopman Lifting

The multi-step EDMD framework generalizes the spectral approach of Koopman operator theory to nonlinear systems, including both deterministic maps and stochastic differential equations (SDEs) with diffusion. Consider a continuous-time SDE

dXt=b(Xt)dt+σ(Xt)dWtdX_t = b(X_t)\,dt + \sigma(X_t)\,dW_t

with initial distribution X0p0(x)X_0 \sim p_0(x), drift b:RdRdb:\mathbb{R}^d\to\mathbb{R}^d, and noise σ:RdRd×s\sigma:\mathbb{R}^d\to\mathbb{R}^{d\times s}, or a discrete-time system xt+1=f(xt,ut)x_{t+1}=f(x_t, u_t). The associated stochastic Koopman semigroup {Kt}\{K^t\} acts on observables ψ\psi by Ktψ(x)=E[ψ(Xtx)]K^t\psi(x) = \mathbb{E}[\psi(X_t^x)], defining a linear semigroup generated by the infinitesimal generator

Lv(x)=b(x)v(x)+12Σ(x):2v(x),Σ(x)=σ(x)σ(x),\mathcal{L}v(x) = b(x)\cdot\nabla v(x) + \frac{1}{2}\Sigma(x):\nabla^2 v(x),\quad \Sigma(x)=\sigma(x)\sigma(x)^\top,

with Fokker–Planck adjoint Lp=[bp]+12[Σp]\mathcal{L}^*p = -\nabla\cdot[b\,p] + \frac{1}{2}\nabla\cdot\nabla\cdot[\Sigma p] (Zhao et al., 2022). Under ergodicity, there exists a unique invariant density p(x)p_*(x), and the L2(Rd;p)L^2(\mathbb{R}^d; p_*) Hilbert space inner product governs the spectral structure of KtK^t and L\mathcal{L}.

Deterministic systems use analogous constructions, with observables {ψj}j=1N\{\psi_j\}_{j=1}^N spanning a finite-dimensional subspace, and evolution in this lifted space governed by approximations of the infinite-dimensional Koopman operator (Schurig et al., 10 Nov 2025, Meda et al., 4 Apr 2025).

2. Finite-Dimensional EDMD Approximations and Multi-Step Forecasting

The core idea of EDMD is to approximate KΔtK^{\Delta t} by a finite-dimensional matrix acting on the lifted state, using empirical data:

  • Snapshot pairs {(xk,yk)}k=1M\{(x_k, y_k)\}_{k=1}^M are generated via dynamics and possible control inputs.
  • The observables define the lift Ψ(x)=[ψ1(x),,ψm(x)]\Psi(x)=[\psi_1(x),\dots,\psi_m(x)]^\top; the data matrices are Ψ(X),Ψ(Y)Rm×M\Psi(X), \Psi(Y)\in\mathbb{R}^{m\times M}.
  • Gram matrices are G=1MΨ(xk)Ψ(xk)G = \frac{1}{M}\sum \Psi(x_k)\Psi(x_k)^\top, A=1MΨ(yk)Ψ(xk)A = \frac{1}{M}\sum \Psi(y_k)\Psi(x_k)^\top.
  • The Koopman approximation is K=Apinv(G)K = A\,\mathrm{pinv}(G), with eigenpairs (μj,ξj)(\mu_j, \xi_j) giving approximate eigenvalues and eigenfunctions (Zhao et al., 2022, Meda et al., 4 Apr 2025).

Multi-step prediction proceeds by recursive iteration:

zk+1=Kzk+Buk,x^k+1=Czk+1,y^k+1=Ezk+1,z_{k+1} = Kz_k + Bu_k,\quad \hat{x}_{k+1} = C z_{k+1},\quad \hat{y}_{k+1} = E z_{k+1},

where the lifted state zk=Ψ(xk)z_k = \Psi(x_k), CC selects physical states from the lifted coordinates, and EE recovers outputs or measurements (Meda et al., 4 Apr 2025). In stochastic density forecasting, spectral decompositions using KK generate multi-step probability evolution via KnK^n or eigenmode propagation (Zhao et al., 2022).

3. Dictionary Design, Manifold Optimization, and Sparsity

The expressiveness and generalization of EDMD surrogates depend critically on the choice of observables. Standard bases include monomials, radial basis functions (RBFs), and neural networks (Meda et al., 4 Apr 2025). Dictionary learning is further refined by geometric optimization on the Grassmann manifold: selecting a subspace V=span{ψ1,,ψM}V=\mathrm{span}\{\psi_1,\dots,\psi_M\} that minimizes multi-step forecast error via Riemannian optimization yields approximately invariant, low-dimensional subspaces that are robust to out-of-domain states (Schurig et al., 10 Nov 2025). The error metric integrates projection error over finite horizons and test initial conditions, and the optimization descends to Grassmannian geometry with Riemannian gradients and QR-based retraction.

Explicit dictionary pruning and structure discovery are enabled by including 1\ell_1-type regularization in least-squares identification, which removes irrelevant observables and yields parsimonious, efficient surrogates (Wu et al., 17 Jan 2026). Parallel decomposition across states, steps, and dictionary rows supports scalable computation.

4. Multi-Step Least-Squares Identification and Error Control

A key advance of the multi-step EDMD framework is direct learning of the condensed multi-step state-output map. Rather than identifying an operator AA for single-step propagation and composing AkA^k (which leads to error compounding and potential instability for A2>1\Vert A\Vert_2>1), the framework fits for each prediction step kk a linear map from dictionary lifts and past inputs to the kk-step-ahead state:

xj,k=Ekψ(xj,0)+Fk[uj,0;;uj,k1]+residual.x_{j,k} = E_k \psi(x_{j,0}) + F_k [u_{j,0};\dots;u_{j,k-1}] + \mathrm{residual}.

The identification is performed via convex least-squares:

minEk,Fkj=1Mxj,kEkψ(xj,0)Fk[uj,0;;uj,k1]22+β[Ek;Fk]22+τEk1,\min_{E_k,F_k} \sum_{j=1}^M \|x_{j,k} - E_k\psi(x_{j,0}) - F_k[u_{j,0};\dots;u_{j,k-1}]\|_2^2 + \beta\|[E_{k};F_{k}]\|_2^2 + \tau\|E_{k}\|_1,

with row-wise 1\ell_1 sparsity for dictionary pruning (Wu et al., 17 Jan 2026). This produces stable, direct multi-step surrogates whose error does not grow with prediction horizon kk. Theorem 3 in (Wu et al., 17 Jan 2026) provides, under Sobolev regularity and bounded-dictionary assumptions, high-probability error bounds independent of kk:

fi(k)Ek,iψHCms(psfi(k)Hs+N2logM/M),\|f_i^{(k)} - E_{k,i}^\top\psi\|_H \leq C_{ms}(p^{-s}\|f_i^{(k)}\|_{H^s} + \sqrt{N^2\log M/M}),

where NN is the dictionary size, MM the number of data, and pp the degree for polynomial lifts. In contrast, one-step EDMD bounds exhibit exponential-in-kk error blow-up when KH>1\|K\|_H > 1.

5. Applications, Numerical Performance, and Practical Considerations

Multi-step EDMD has been demonstrated for prediction, nonlinear system identification, density forecasting, and model predictive control (MPC) across a wide range of systems:

  • Density forecast for SDEs: Approximates the semigroup generated by the Fokker–Planck operator, achieving weak convergence of the density representation as data and dictionary sizes increase (Zhao et al., 2022).
  • Nonlinear system surrogates: For electric vehicle (EV) cabin climate, multi-step EDMD with RBF dictionaries achieves state RMSE <5%<5\% and power RMSE <3%<3\% for N=35N=35 basis, outperforming polynomial and neural-network dictionaries at small NN (Meda et al., 4 Apr 2025).
  • Koopman MPC: Direct multi-step identification solves the full condensed map for state and control over control horizon TT, enabling stable closed-loop control where one-step EDMD leads to diverging errors and controller instability. In benchmarks, multi-step EDMD achieves stable MSE under open and closed loop for k10k\gg 10, while one-step MSE diverges for k>10k>10 due to spectral radius >1>1 (Wu et al., 17 Jan 2026).
  • Dictionary learning: Grassmannian-based shaping yields low-dimensional surrogates with out-of-domain generalization and large speedups in predictor evaluation (Schurig et al., 10 Nov 2025).

Per-step computational costs are governed by O(N2)O(N^2) matrix-vector products and O(Nϕ)O(N\phi) dictionary evaluation (with ϕ\phi cost per basis). For N30N\sim 30–$40$, real-time execution on embedded platforms is practical (Meda et al., 4 Apr 2025).

6. Algorithmic and Implementation Summary

A generic multi-step EDMD workflow proceeds as:

  1. Dictionary selection: Choose functions {ψj}\{\psi_j\}—polynomial, RBF, or learned—appropriate to system structure and data availability.
  2. Data acquisition: Collect snapshot (state, next-state) or (trajectory) data under representative state and control distributions.
  3. Lifting and matrix assembly: Formulate Ψ(X)\Psi(X), Ψ(Y)\Psi(Y) and associated Gram matrices; assemble multi-step input-output structures for supervised learning.
  4. Identification: Solve (regularized) least-squares for operator matrices, possibly under parallelizable decomposition across step kk and state ii; perform dictionary pruning as necessary.
  5. Prediction: For given initial condition and input sequence, propagate via

zk+1=Kzk+Buk,orxk=Ekψ(x0)+Fk[u0;;uk1]z_{k+1} = K z_k + B u_k\,,\quad\text{or}\quad x_{k} = E_k \psi(x_0) + F_k [u_0;\ldots;u_{k-1}]

up to desired horizon, reconstruct state and output via CC and EE matrices.

  1. Validation: Assess with trajectory RMSE, Consistency Index (CI), or finite-horizon forecast errors; compare performance with alternative dictionaries or identification schemes (Zhao et al., 2022, Meda et al., 4 Apr 2025, Schurig et al., 10 Nov 2025, Wu et al., 17 Jan 2026).

When relevant (e.g., Grassmannian dictionary learning), implement Riemannian optimization to shape active observables for improved multi-step invariance and forecast accuracy (Schurig et al., 10 Nov 2025). Closed-loop extensions with linear MPC/LQR or online adaptation using recursive least squares are direct, leveraging the lifted linear structure of the surrogate.

7. Convergence Guarantees and Limitations

Convergence properties are established in terms of increasing data (MM\to\infty), basis set expressiveness (m,Nm,N\to\infty), and regularity of the lifting functions. For stochastic systems, EDMD approximations converge to the Galerkin projection of the Fokker–Planck semigroup in L2(p)L^2(p_\ast), and the truncated density forecast pm(x,t)p_m(x,t) converges weakly to the true solution (Zhao et al., 2022). In deterministic and control settings, multi-step EDMD avoids error blow-up inherent to one-step recursion, with finite-time sample error scaling optimally in dictionary size and sample count (Wu et al., 17 Jan 2026). Overfitting and lack of generalization can arise with excessive dictionary dimension, motivating use of Grassmannian shaping and sparsity regularization (Schurig et al., 10 Nov 2025, Wu et al., 17 Jan 2026).

A plausible implication is that careful system-specific dictionary design, appropriate regularization, and direct multi-horizon identification are essential for the robust practical application of EDMD in high-dimensional settings and real-time control scenarios.

Topic to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Multi-step EDMD Framework.