Conformal Online Koopman Learning (COLoKe)
- The paper introduces a conformal update mechanism that selectively refines model parameters based on deviations in multi-step prediction accuracy.
- It integrates deep neural network–based embeddings with a temporal consistency loss to lift nonlinear dynamics into a linear, high-dimensional space.
- Empirical results on synthetic and real-world benchmarks demonstrate orders-of-magnitude improvements in prediction accuracy with reduced update frequency.
Conformal Online Learning of Koopman Embeddings (COLoKe) is a framework for adaptively learning Koopman-invariant representations of nonlinear dynamical systems from streaming data. By integrating deep neural network–based embeddings with a conformal-style, event-triggered control mechanism, COLoKe ensures that the learning of both the lifting map and the linear Koopman operator is performed both efficiently and with explicit regularization to avoid overfitting. The critical innovation is a selective update scheme: model parameters are refined only when the predicted conformity of the current Koopman model falls below a dynamically calibrated threshold, as measured over a rolling temporal window. This enables sustained predictive accuracy with reduced (and adaptive) update frequency while maintaining provable dynamic-regret guarantees and minimizing unnecessary computation (Gao et al., 16 Nov 2025).
1. Model Architecture and Koopman Embedding
COLoKe lifts each raw system state into a higher-dimensional complex vector via
where is parameterized by a neural network with trainable weights . The explicit inclusion of the identity component in the output enforces a built-in reconstruction constraint, ensuring that original system states are preserved in the embedding without a separate decoder.
In this lifted space, COLoKe seeks a Koopman operator such that the dynamics propagate approximately linearly: and for multistep prediction: This approach generalizes standard deep Koopman operator learning to the online setting, where both and are updated as new data streams in.
2. Multistep Temporal Consistency and Loss Formulation
To enforce temporal consistency, COLoKe maintains a buffer of the latest states. The multi-step consistency loss at time is defined by
where . This loss enforces that the learned embedding and operator jointly produce consistent multi-step predictions across all temporally local subwindows.
At every time , the conformity of the current model is assessed using the prediction error at the buffer's end: This scalar score represents the discrepancy between multi-step forward predictions and actual states, serving as the central criterion for triggering model updates.
3. Conformal-Style Update Trigger and Threshold Control
Rather than updating model weights at every time step, COLoKe utilizes a conformal-style mechanism that adaptively controls when updates occur. A nonconformity threshold is maintained and adjusted by a proportional–integral (PI) control law: where is an indicator for excess prediction error, is the target nonconformity rate, is a step-size parameter, and is a saturation constant for the integral term to stabilize control.
Model updates are triggered only if . When this occurs, gradient descent steps are performed on (over both and ) until . This selective adaptation provides strong regularization: updates are explicit responses to statistically significant model misfit, rather than being performed indiscriminately.
The update rule can be summarized as:
- If : set ;
- Else: perform gradient steps on until .
A key property is that updating terminates as soon as the conformity test is satisfied, which acts as a natural safeguard against overfitting.
4. Online Algorithm Workflow
A concise description of the learning workflow is as follows:
- Initialize embedding parameters and operator via fitting on the first data points. Set initial to the -quantile of the initial conformity scores.
- For each new time step :
- Observe and update buffer ;
- Copy ;
- Compute conformity score ;
- Update the threshold using the PI control law;
- If , perform gradient steps on and recompute until ;
- Advance to the next time step.
Stopping update iterations as soon as conformity is restored avoids over-minimization, further supporting generalization.
5. Empirical Evaluation and Performance
COLoKe was benchmarked on classical synthetic dynamical systems—single attractor systems (with analytic spectra), the Duffing oscillator (mixed spirals and saddle points), the Van der Pol oscillator (limit cycle), and the Lorenz system (chaotic attractor)—as well as real-world time series (electricity transformer, EEG, atmospheric turbulence) (Gao et al., 16 Nov 2025).
The primary evaluation metrics used are:
- Generalization error: one-step MSE on held-out trajectories;
- Online prediction error: average MSE as streaming data is processed;
- Update frequency: the fraction of steps where .
Across these benchmarks, COLoKe consistently achieves:
- The lowest prediction errors, with orders-of-magnitude improvement on the Lorenz system;
- Approximately 50% update frequency, representing significantly fewer updates than fixed-step baselines (which update 100% of steps);
- Accurate recovery of Koopman eigenvalues and eigenfunctions;
- Reduced wall-clock training time compared to offline deep Koopman approaches.
A plausible implication is that conformal-style selective updating provides a favorable tradeoff between prediction accuracy and computational effort in online dynamical system identification.
6. Hyperparameter Choices and Sensitivity
The main hyperparameters and their effects are summarized below:
| Hyperparameter | Typical Range | Effect |
|---|---|---|
| Window size | 10–50 | Controls horizon of multistep loss; larger enforces greater temporal consistency but increases per-step computation |
| Conformal PI | High : higher accuracy, more updates; low : fewer updates, possible degradation in fit | |
| PI step | Adjusts agility of threshold adaptation | |
| I-term saturation | bounded by | Ensures stability of PI controller |
| Learning rate | e.g., | Sets step size in triggered gradient updates |
| Embedding dimension | Higher improves expressivity at the cost of parameter count |
These settings enable COLoKe to adapt to both fast-changing and stable regimes, maintain computational efficiency, and control overfitting via principled, data-dependent update rules.
7. Principles and Broader Relevance
COLoKe exemplifies an approach where learning of deep operator-theoretic models is embedded within a conformal-style supervisory control loop. This separates the questions of how to learn (via deep neural parameterization and multistep temporal consistency) from when to update (decided by online conformity scores relative to a statistically motivated, dynamically calibrated threshold). The method yields provably efficient, online adaptive learning with dynamic-regret guarantees and avoids pitfalls of overfitting endemic to naive online retraining. Demonstrated empirical success on synthetic and real benchmarks highlights its practicality for streaming and nonstationary system identification scenarios (Gao et al., 16 Nov 2025).