wtSINDy: Weighted Sparse ID of Nonlinear Dynamics
- wtSINDy is a data-driven method that refines traditional sparse identification by introducing non-uniform weights to enhance convergence in dynamic operator estimation.
- It mitigates edge effects and filters high-frequency noise, improving estimation accuracy across periodic, quasiperiodic, and stochastic regimes.
- The approach leverages weighted Birkhoff averages, clustering, and regularized regression techniques to achieve robust performance and reduced data requirements.
Weighted Extended Dynamic Mode Decomposition (wtEDMD) is a data-driven algorithm for extracting spectral properties, reduced-order models, and time-evolution operators associated with ergodic dynamical systems and stochastic differential equations. By introducing non-uniform, smoothly vanishing weights into the standard extended DMD framework, wtEDMD markedly accelerates convergence of finite-data estimates, especially in periodic and quasiperiodic regimes, and enhances filtering of high-frequency noise in stochastic systems. Rigorous convergence rates and practical algorithmic steps are detailed in the works of Bramburger, Colbrook, and Tahara et al. (Bou-Sakr-El-Tayar et al., 21 Nov 2025, Tahara et al., 26 Mar 2024).
1. Mathematical Foundations
In wtEDMD, core ergodic averages underpin operator estimation. For a measure-preserving, ergodic system and observable , the classic Birkhoff average is
wtEDMD replaces the uniform weights by a smooth taper satisfying , , and all derivatives vanishing at endpoints,
A canonical bump function is , with normalization constant set by . These weights inherently suppress edge effects and facilitate super-polynomial or exponential convergence in non-chaotic settings (Bou-Sakr-El-Tayar et al., 21 Nov 2025).
In the stochastic context (see (Tahara et al., 26 Mar 2024)), consider the Itô SDE
with drift , diffusion matrix , and associated generator
where . Locally weighted expectation operators are introduced to robustly estimate such derivatives,
with Gaussian kernel .
2. wtEDMD Algorithm and Regression System
The standard EDMD constructs least-squares approximations of the Koopman operator using uniform averages over observables. wtEDMD modifies the approach by introducing a diagonal weight matrix composed of ,
and solves
yielding the optimal (Bou-Sakr-El-Tayar et al., 21 Nov 2025).
Analogously, for the Koopman generator, the regression system is set up via weighted conditional moments at representative points,
where encodes generator derivatives estimated from weighted local statistics, and is the finite-dimensional Koopman generator (Tahara et al., 26 Mar 2024).
3. Convergence Properties
Weighted Birkhoff averages in wtEDMD exhibit markedly improved convergence rates compared to uniform averaging:
- Periodic dynamics: exponential decay in .
- Quasiperiodic dynamics with smooth observables: super-polynomial error for all .
- Analytic observables on quasiperiodic orbits: exponential convergence.
- Chaotic or stochastic settings: empirically at least , matching conventional EDMD rates (Bou-Sakr-El-Tayar et al., 21 Nov 2025).
These improvements are retained in finite-data projections of the Koopman operator and generator. Operator-norm bounds follow via weighted ergodic theory and convergence results for EDMD.
4. Practical Implementation and Clustering
In practice, wtEDMD requires careful selection of weights, basis size, and sample counts for stability:
- Typical weights: bump functions, or signal-processing windows (e.g., Tukey, Blackman–Harris) adjusted to kill endpoint derivatives.
- Sample count : must exceed basis sizes for stable regression; in regular regimes, can often be reduced.
- Conditioning: monitor ; if ill-conditioned, apply Tikhonov regularization or use stable SVD pseudoinverse.
For stochastic systems, Tahara et al. (Tahara et al., 26 Mar 2024) advocate a two-stage clustering methodology:
- Outlier Removal: IsolationForest eliminates anomalous trajectory points.
- Representative Selection: k-means yields centroids to focus the regression.
- Local Structure Capture: Dirichlet-process Gaussian-mixture model (DPMM) clusters the dataset, with each centroid inheriting its cluster covariance for anisotropic weighting.
This enables local moments and generator derivatives to reflect trajectory-dependent noise geometry, yielding substantially better drift/diffusion estimates and Koopman generator matrices.
5. Numerical Illustrations
To clarify performance gains and practical trade-offs, select benchmarks are summarized in the table below.
| Example & Regime | EDMD Error | wtEDMD Error |
|---|---|---|
| Standard map, quasiperiodic (), 9×9 Fourier | () | , plateaus at precision |
| Lid-driven cavity, autocorrelation (periodic) | () | (, weighted) |
| El Niño diffusion forecast, RMSE/correlation (16 mo lead) | Baseline | wtEDMD: correlation gain ∼0.1 |
A plausible implication is that, in non-chaotic settings, wtEDMD can reduce data requirements and improve estimation accuracy by several orders of magnitude. For chaotic systems, wtEDMD matches EDMD rates and does not introduce slowdown.
6. Limitations and Best Practices
wtEDMD's efficacy is regime-dependent. In chaotic or strongly stochastic dynamics, speed-up is limited to . Noisy or insufficiently smooth observables can be amplified at endpoints if heavy tapering is applied—moderate windows or pre-filtering may be preferred. Overcomplete basis sets necessitate significantly larger data sample sizes to avoid ill-conditioning in the weighted pseudocovariance matrices.
Weight profile tuning balances edge suppression and sample efficiency: strong damping loses sample size; 10–20% Tukey windows provide a compromise. Always use robust SVD-based pseudoinverses and regularize Gram matrices when required.
7. Contextual Significance
wtEDMD represents a refinement of time-averaged data-driven operator estimation found in DMD, EDMD, SINDy, spectral measure estimation, and diffusion forecasting. By integrating weighted Birkhoff theory and local weighted conditional expectation (in the stochastic generator context), the method directly addresses finite-data limitations and improves convergence without additional coding overhead. Its two-stage clustering for stochastic systems reflects recent advances in capturing anisotropic noise via unsupervised learning priors, contributing to improved generator estimation in nonlinear systems (Bou-Sakr-El-Tayar et al., 21 Nov 2025, Tahara et al., 26 Mar 2024).