Koopman Operator Prediction
- Koopman operator prediction is a mathematical framework that transforms nonlinear system dynamics into a linear evolution through data-driven eigenfunction construction.
- It leverages spectral properties and convex optimization to lift state variables into high-dimensional observable spaces for precise trajectory and state estimation.
- The approach is applicable to control, turbulence modeling, and system identification, integrating theoretical density guarantees with practical model predictive control.
Koopman operator prediction is a mathematical, computational, and data-driven framework for forecasting the evolution of nonlinear dynamical systems by employing a linear operator acting on an (often infinite-dimensional) space of observables. This approach leverages the spectral properties of the Koopman operator to construct linear models that, when combined with appropriate lifting functions (eigenfunctions or learned embeddings), can accurately predict both trajectories and statistical properties of complex systems. Koopman-based prediction has found applications in system identification, control, turbulence modeling, uncertainty quantification, epidemic forecasting, and nonlinear state estimation, and underpins several modern advances in machine learning and data-driven modeling of nonlinear phenomena.
1. Koopman Eigenfunction Construction and Linearization
The construction of Koopman eigenfunctions is central to casting nonlinear system evolution into a linear framework. The data-driven methodology introduced by Korda and Mezić (Korda et al., 2018) eschews fixed dictionary selection by defining Koopman eigenfunctions as follows. Given a non-recurrent set Γ and a collection of continuous boundary functions , for each candidate eigenvalue the eigenfunction on the flow-generated domain is constructed as
where is the (backward) hitting time of along the flow . This construction solves the differential equation
with initial condition on . The richness of the Koopman spectrum away from attractors means that, via suitable optimization over and , a dense set of eigenfunctions can be generated. The mapping is linear in , allowing convex optimization (e.g., -regularized least squares) to minimize projection error of observable quantities onto the eigenfunction span. The selection of is non-convex but tractable, and can be optimized via gradient-based methods.
Generalized eigenfunctions associated with Jordan blocks can similarly be constructed: yielding richer invariant subspaces and enabling linear prediction even under spectral degeneracy.
2. Data-Driven Prediction and Control: Linear Lifting and MPC
After constructing a set of Koopman eigenfunctions , the state is lifted to , leading to an exactly linear evolution in the absence of control: where is obtained by solving an optimization (typically quadratic) to minimize the error in approximating the observable of interest.
For control-affine systems , the lifted dynamics become
Here, and are fixed from the uncontrolled system, and is fitted using convex multi-step least squares on data with nonzero input, minimizing the multi-step prediction error. The predictor structure, determined entirely by data-driven convex optimization, is then exploited in a model predictive control (MPC) framework: the finite-horizon optimal control problem is formulated as a convex quadratic program over the lifted system, allowing direct application of linear MPC tools for controlling nonlinear dynamics.
3. Theoretical Guarantees: Density, Generalized Eigenfunctions, and Invariant Subspaces
The foundational result (Theorem 3.1 in (Korda et al., 2018)) states that, for any set (containing at least one eigenvalue with nonzero real part) and a countable, dense family , the span of the generated eigenfunctions becomes dense in . Thus, every continuous observable in a sufficiently reachable domain can be approximated arbitrarily well in the Koopman invariant subspace generated via this construction.
The extension to generalized eigenfunctions—via chains parametrized by Jordan matrices —allows for the construction of invariant subspaces even in the presence of non-trivial algebraic or geometric multiplicity, thereby supporting linear prediction, spectral analysis, and model reduction tasks.
4. Comparative Numerical Results, Prediction Accuracy, and Feedback Control
Numerical examples in (Korda et al., 2018) demonstrate the practical efficacy of the eigenfunction-based linear predictors in both prediction and control:
- Van der Pol oscillator: Lifting from a non-recurrent set yields accurate state prediction; the error is minimized via eigenvalue optimization.
- Damped Duffing oscillator: The framework captures multi-modal equilibrium structure, phase transitions (e.g., stabilization at unstable equilibria via Koopman MPC), and enables successful tracking between equilibrium points.
Performance is compared between predictors using “as-is” eigenvalues (e.g., from DMD) and those using optimized eigenvalues (via nonconvex, gradient-based search), with the latter showing substantially lower prediction errors. Closed-loop control experiments on the Duffing system illustrate effective MPC deployment, with solution times and closed-loop behavior reflecting both computational efficiency and feedback robustness.
Moreover, the modularity of the construction permits post hoc enrichment of the predictor basis (by exploiting the algebraic closure of eigenfunctions under products and powers) without recomputation of the entire model. All examples are accompanied by publicly released Matlab code, ensuring reproducibility and accessibility.
5. Structural and Algorithmic Features
Key attributes of the framework are:
- Dictionary-free construction: No fixed set of basis functions is required; eigenfunctions are defined via boundary data and optimized globally.
- Convexity: All stages (except eigenvalue search) are cast as convex optimizations — specifically, linear/quadratic programs or convex least squares.
- Generalizability: The architecture extends naturally to controlled settings, multi-step prediction, and model predictive control.
- Modularity: Eigenfunctions for additional observables or interpolation at new points can be generated on demand, reflecting the algebraic structure of the Koopman operator.
- Theoretical rigor: Density results guarantee that the constructed model can, in principle, capture any target observable up to required fidelity.
6. Limitations and Deployment Considerations
Potential limitations include:
- Dependence on non-recurrent set selection: The quality and reachability of the non-recurrent set affects coverage in state space.
- Necessity of sufficient data: Approximating the Koopman invariant subspace requires adequate excitation and coverage via data.
- Choice of eigenvalues: While the optimization of is low-dimensional and tractable, the non-convexity can potentially lead to sub-optimal local minima.
- Scalability: For high-dimensional systems, computational cost scales with the number of eigenfunctions and the complexity of the optimization landscape.
- Applicability near attractors: The approach leverages spectral richness away from attractors; on or near attractors the spectrum condenses and richer sets of eigenfunctions may need to be computed.
Deployment in real-world settings benefits from the framework’s convexity and dictionary-free design, which avoid issues of basis selection and non-convex neural network training. The method is directly compatible with existing convex optimization libraries and can be extended to larger systems given sufficient computational resources.
7. Impact and Integration with Broader Koopman Operator Research
The explicit, optimization-driven construction of eigenfunctions and prediction models offers a principled bridge between theoretical guarantees (density of eigenfunction spans) and practical, modular architectures for high-accuracy, data-driven prediction and control in nonlinear systems. By focusing on convexity, modularity, and post hoc extensibility, this method stands in contrast to black-box, non-convex machine learning techniques or heavily parametric approaches, and offers a robust foundation for ongoing developments in Koopman-based prediction, cavity model reduction, and data-driven control synthesis. The release of accompanying code further solidifies its position as a reference implementation in the field.