Weighted Extended DMD (wtEDMD) Overview
- wtEDMD is a data-driven method that replaces uniform ergodic averages with smooth weighted profiles, enhancing the accuracy and convergence of Koopman operator approximations.
- It integrates specialized clustering and local weighting techniques to accurately estimate drift and diffusion in nonlinear stochastic and deterministic systems.
- The algorithm exhibits faster convergence rates for periodic and quasiperiodic dynamics while maintaining robust performance in chaotic or noisy settings.
Weighted Extended Dynamic Mode Decomposition (wtEDMD) is a data-driven algorithm for operator approximation in dynamical systems that leverages weighted ergodic averages to suppress edge effects and accelerate convergence relative to standard Extended Dynamic Mode Decomposition (EDMD). In wtEDMD, smooth, vanishing-endpoint weight profiles replace uniform averaging, and specialized clustering and local weighting techniques enable robust performance even in nonlinear stochastic systems. The approach encompasses both finite-time Koopman operator approximations and Koopman generator estimation for stochastic differential equations, providing spectral analysis and system identification from limited or noisy data (Bou-Sakr-El-Tayar et al., 21 Nov 2025, Tahara et al., 26 Mar 2024).
1. Mathematical Foundations of Weighted Averages
wtEDMD modifies the classical ergodic Birkhoff average by introducing a smooth, non-uniform weight function. Given a measure-preserving ergodic system and observable , the standard average is
wtEDMD replaces this with
where satisfies and for all (Bou-Sakr-El-Tayar et al., 21 Nov 2025). Common choices include bump functions .
For stochastic systems, wtEDMD leverages locally weighted expectation operators. Conditional averages approximating the drift and diffusion— and , respectively—are computed via kernels defined by the local sample covariance (Tahara et al., 26 Mar 2024).
2. Algorithmic Structures and Recipes
wtEDMD for Koopman Operator
The algorithm proceeds as follows (Bou-Sakr-El-Tayar et al., 21 Nov 2025):
- Snapshot Collection: Gather sequential snapshots .
- Dictionary Construction: Define observable dictionaries , .
- Weight Matrix: Build .
- Weighted Least Squares: Solve
yielding , where and are matrices of observables.
- Spectral Analysis: serves as the projected Koopman operator.
wtEDMD for Koopman Generator (Stochastic SDEs)
For Itô SDEs, wtEDMD first computes locally weighted conditional moments (Tahara et al., 26 Mar 2024):
- Estimate drift and diffusion at representative points via kernels with cluster-specific bandwidths.
- Construct derivative-observable pairs at centroids .
- Solve a weighted (optionally sparse) least-squares or lasso regression problem for the generator approximation matrix .
Pipeline Summary for Clustered wtEDMD
| Step | Method | Purpose |
|---|---|---|
| Outlier removal | IsolationForest | Discard sparse/noisy points |
| Rep. selection | k-means clustering | Reduce to centroids |
| Local structure | Dirichlet Process Gaussian Mixture (DPMM) | Capture anisotropy for local weighting |
3. Convergence Properties and Error Rates
Weighted Birkhoff averages enable faster convergence than uniform averaging, as edge effects from finite data are suppressed (Bou-Sakr-El-Tayar et al., 21 Nov 2025). Proven rates:
- Periodic : Exponential decay in .
- Quasiperiodic, smooth : Super-polynomial error for all .
- Analytic , quasiperiodic orbits: Exponential decay.
- Chaotic/stochastic: Empirical decay matches classical rates; wtEDMD does not degrade convergence.
Replacing the averages in EDMD with weighted averages retains asymptotic operator limits but accelerates practical convergence, especially for regular (periodic, quasiperiodic) dynamics (Bou-Sakr-El-Tayar et al., 21 Nov 2025). In stochastic settings, localized expectation operators filter high-frequency noise, yielding generator matrices more accurate than unweighted or naive finite-difference approaches (Tahara et al., 26 Mar 2024).
4. Practical Implementation and Numerical Examples
Key considerations for stable implementation (Bou-Sakr-El-Tayar et al., 21 Nov 2025, Tahara et al., 26 Mar 2024):
- Weight functions: Prefer bump functions or modified signal-processing windows with vanishing endpoint derivatives.
- Sample sizes: should exceed the dictionary sizes substantially for stability.
- Conditioning: Monitor pseudocovariances; apply Tikhonov regularization or SVD-based pseudoinverse if ill-conditioned.
- Window tuning: Aggressive tapering suppresses edge artifacts but reduces effective sample size near boundaries.
Representative Results
| Example | Standard EDMD Error | wtEDMD Error / Convergence | Dynamics |
|---|---|---|---|
| Standard map (Fourier dict.) | ; superpoly. | Quasiperiodic | |
| Standard map () | Chaotic | ||
| Lid-driven cavity autocomp. | Periodic | ||
| El Niño forecast (RMSE) | Uniformly improved | modes optimal | Stochastic |
In the El Niño diffusion forecast, weighted forecasting improved RMSE and correlation for leads up to 16 months, but excessive mode counts () with limited can degrade performance (Bou-Sakr-El-Tayar et al., 21 Nov 2025).
5. Limitations and Best Practices
- Stochastic/chaotic systems: wtEDMD’s speedup is limited to , since ergodic averages' convergence is fundamentally statistical (Bou-Sakr-El-Tayar et al., 21 Nov 2025).
- Noise/smoothness requirements: Strong tapers may amplify endpoint noise—apply milder windows or pre-filter data in such cases.
- Dictionary size: Avoid overcomplete bases unless .
- Weight tuning: Balance endpoint suppression and effective sample size; compromise with partial Tukey windows and monitor normalization constant .
- Numerical stability: SVD-based or regularized pseudoinverses are essential.
For stochastic systems, cluster-specific kernels (DPMM bandwidth matrices) outperform isotropic weighting, and two-stage clustering ensures representative diversity and local statistical consistency (Tahara et al., 26 Mar 2024).
6. Connections and Extensions
wtEDMD generalizes the weighted ergodic averaging principle to a range of operator-theoretic and identification algorithms, including wtDMD, wtSINDy, weighted spectral measure estimation, and diffusion forecasting (Bou-Sakr-El-Tayar et al., 21 Nov 2025). For generator estimation in SDEs, locally weighted expectations with clustering improve drift/diffusion reconstruction, outperforming classical regression and naive kernel methods (Tahara et al., 26 Mar 2024). Future work may address rigorously quantifying error propagation for chaotic systems and exploring adaptive weight profiles for systems with highly nonuniform sampling or strong state dependence.
The methodology is extensible to other data-driven frameworks relying on time-averaged statistics and is compatible with regularization and dimensionality reduction techniques. A plausible implication is that weighting and clustering can be systematically combined for improved system identification in the presence of non-uniform noise, anisotropic uncertainty, or limited samples.