Papers
Topics
Authors
Recent
Search
2000 character limit reached

HyperKKL: Learning KKL Observers for Non-Autonomous Nonlinear Systems via Hypernetwork-Based Input Conditioning

Published 31 Mar 2026 in eess.SY and cs.LG | (2603.29744v1)

Abstract: Kazantzis-Kravaris/Luenberger (KKL) observers are a class of state observers for nonlinear systems that rely on an injective map to transform the nonlinear dynamics into a stable quasi-linear latent space, from where the state estimate is obtained in the original coordinates via a left inverse of the transformation map. Current learning-based methods for these maps are designed exclusively for autonomous systems and do not generalize well to controlled or non-autonomous systems. In this paper, we propose two learning-based designs of neural KKL observers for non-autonomous systems whose dynamics are influenced by exogenous inputs. To this end, a hypernetwork-based framework ($HyperKKL$) is proposed with two input-conditioning strategies. First, an augmented observer approach ($HyperKKL_{obs}$) adds input-dependent corrections to the latent observer dynamics while retaining static transformation maps. Second, a dynamic observer approach ($HyperKKL_{dyn}$) employs a hypernetwork to generate encoder and decoder weights that are input-dependent, yielding time-varying transformation maps. We derive a theoretical worst-case bound on the state estimation error. Numerical evaluations on four nonlinear benchmark systems show that input conditioning yields consistent improvements in estimation accuracy over static autonomous maps, with an average symmetric mean absolute percentage error (SMAPE) reduction of 29% across all non-zero input regimes.

Summary

  • The paper introduces HyperKKL, a framework that applies hypernetwork-based input conditioning to yield input-dependent neural KKL observers for improved state estimation in non-autonomous nonlinear systems.
  • The framework comprises two strategies, HyperKKL_obs and HyperKKL_dyn, which adjust latent dynamics and encoder-decoder weights respectively, achieving up to 86% SMAPE reduction over static baselines.
  • The study provides theoretical error bounds and practical guidelines, highlighting a trade-off between dynamic expressiveness and optimization complexity for robust observer design.

HyperKKL: Hypernetwork-Based Input Conditioning for Non-Autonomous KKL Observers

Overview and Motivation

The paper investigates the synthesis of Kazantzis-Kravaris/Luenberger (KKL) observers for nonlinear non-autonomous systems, specifically those whose dynamics are influenced by exogenous inputs. Traditional neural KKL observer architectures are restricted to autonomous regimes, relying on static transformation maps that do not adequately generalize to input-driven systems. This architectural limitation results in persistent phase drift and degraded estimation accuracy when applied to controlled systems. Theoretical extensions in prior work suggest that time-varying transformation maps are necessary to accommodate non-autonomous settings, but systematic learning-based implementations for such observers have been lacking.

To bridge this gap, the authors propose the HyperKKL framework, leveraging hypernetwork-based input conditioning for neural KKL observers. Two complementary strategies are introduced: (i) HyperKKL\textsubscript{obs}, which applies input-dependent corrections to latent observer dynamics while retaining static encoder-decoder maps; and (ii) HyperKKL\textsubscript{dyn}, which uses a hypernetwork to generate input-conditioned encoder and decoder weights for fully dynamic transformation maps. The framework achieves significant improvements in state estimation error, especially across diverse input regimes, and provides explicit theoretical bounds relating approximation and architectural properties to estimation performance.

HyperKKL Framework: Architectural Innovations

HyperKKL is structured in two phases: autonomous pretraining and input-conditioned adaptation.

  • Phase 1—Autonomous Pretraining: Encoder and decoder networks are separately trained on data from the unforced (autonomous) system. The encoder approximates a transformation map T\mathcal{T} into a stable latent space, and the decoder approximates the left inverse T\mathcal{T}^*.
  • Phase 2—Input Conditioning:
    • HyperKKL\textsubscript{obs}: A gated recurrent unit (GRU) processes the input history window to generate a temporal context embedding. An MLP injection network then maps this embedding and latent state to input-dependent corrections Φ\Phi in the observer dynamics; encoder and decoder remain frozen.
    • HyperKKL\textsubscript{dyn}: A hypernetwork generates input-dependent weight perturbations for each layer of the encoder and decoder, enabling dynamic transformation and inversion conditioned on recent input history. This architecture is more expressive but also introduces optimization complexity.
    • Figure 1
    • Figure 1: HyperKKL\text{HyperKKL} schematic showing input-conditioned weight modulation via the hypernetwork for encoder and decoder.

Both strategies are trained using a dataset comprising diverse initial conditions and input signals, enabling generalization across multiple regimes.

Theoretical Guarantees: Estimation Error Bounds

The authors derive a worst-case bound for the state estimation error of the learned KKL observer. The error is shown to depend on: (1) the uniform approximation error of the learned encoder and decoder; (2) the residual violation of the governing PDE for the transformation map; and (3) the Lipschitz constant of the decoder inversion. Explicitly, for compact domains and continuously differentiable networks, the asymptotic estimator error is bounded by:

lim suptξ(t)ϵdec+dec(ϵpdeκλ+ϵenc)\limsup_{t\to\infty} \|\xi(t)\| \leq \epsilon_{\text{dec}} + \ell_{\text{dec}} \left( \epsilon_{\text{pde}}\frac{\kappa}{\lambda} + \epsilon_{\text{enc}} \right)

where ϵpde\epsilon_{\text{pde}} is the maximal PDE residual, dec\ell_{\text{dec}} is the decoder's Lipschitz constant, ϵenc\epsilon_{\text{enc}}, ϵdec\epsilon_{\text{dec}} are encoder and decoder approximation errors, and κ\kappa, T\mathcal{T}^*0 arise from the Hurwitz observer matrix. The architectural difference between HyperKKL\textsubscript{obs} and HyperKKL\textsubscript{dyn} manifests in their control over T\mathcal{T}^*1 and T\mathcal{T}^*2, explaining the observed empirical trade-offs.

Empirical Validation: Robust Accuracy Gains

The framework is tested on four benchmark nonlinear systems (Duffing, Van der Pol, Rössler, FitzHugh-Nagumo) under various input regimes (zero, constant, sinusoidal, square wave). Both observer variants are compared to baselines: the static autonomous observer and a curriculum learning strategy. Strong numerical results include:

  • SMAPE reduction: HyperKKL\textsubscript{obs} achieves up to 86% reduction in symmetric mean absolute percentage error (SMAPE) compared to the autonomous baseline under certain input regimes.
  • Consistency: Both HyperKKL variants reduce SMAPE under non-zero input in nearly all cases, with HyperKKL\textsubscript{obs} outperforming HyperKKL\textsubscript{dyn} in 13 of 16 system/input combinations.
  • Generalization: The curriculum learning baseline, which relies only on diverse data, exhibits degraded performance, confirming that architectural input conditioning—not richer data alone—is necessary for robust observer design. Figure 2

Figure 2

Figure 2

Figure 2

Figure 2

Figure 2: State estimation of the Duffing oscillator under square wave input, demonstrating that HyperKKL\textsubscript{obs} accurately tracks the ground truth and avoids phase drift.

Practical and Theoretical Implications

The findings have several direct implications:

  • Input conditioning is essential: Static transformation maps are insufficient for non-autonomous nonlinear systems; architectural integration of input signals into observer dynamics or mapping is required to ensure bounded error and correct embedding.
  • Hypernetworks as adaptation mechanisms: Hypernetwork-based conditioning enables interpolation in the space of transformation maps, supporting robust generalization across input regimes without retraining for each realization.
  • Optimization/expressiveness trade-off: Dynamic conditioning (HyperKKL\textsubscript{dyn}) is theoretically more expressive, but practical optimization difficulties mean latent injection (HyperKKL\textsubscript{obs}) often offers better accuracy.
  • Future directions: The analysis motivates spectral normalization to control the Lipschitz constant, hybrid approaches focusing on decoder conditioning, and disturbance-aware extensions.

Conclusion

HyperKKL establishes a systematic hypernetwork-based approach to learning input-conditioned neural KKL observers for non-autonomous nonlinear systems. Both latent injection and dynamic transformation conditioning methods yield substantial reductions in estimation error compared to autonomous observers, explicitly addressing the architectural gap at the intersection of system identification and neural observer design. Theoretical analysis clarifies the underlying trade-offs and guides future model and optimization improvements, with broad implications for robust nonlinear state estimation and adaptive observer synthesis in complex, input-driven environments.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.