Papers
Topics
Authors
Recent
2000 character limit reached

Augmented Kalman Filter (AKF)

Updated 29 November 2025
  • Augmented Kalman Filter (AKF) is an advanced state estimation method that jointly identifies model parameters and induces sparsity for adaptive, online system identification.
  • It utilizes an augmented state-space model with UKF-based prediction-update steps and incorporates Bayesian ARD for dynamic, interpretable model structure selection.
  • AKF demonstrates practical gains in control engineering and computational physics, offering improved accuracy and robustness over standard filtering approaches.

The Augmented Kalman Filter (AKF) is an extension of classic Kalman filtering frameworks that incorporates model-parameter identification and sparse model structure selection jointly with state estimation. This paradigm is foundational for online dynamic system identification under noise, partial observability, and model uncertainty. AKF serves as the backbone for a variety of Sparse Kalman Identification (SKI) algorithms, all of which leverage recursive Bayesian filtering and sparsification strategies for interpretable, adaptive, and accurate physical modeling in fields ranging from control engineering to computational physics (Mei et al., 22 Nov 2025).

1. Mathematical Formulation and State–Parameter Augmentation

AKF operates on a discrete-time nonlinear state-space system: xt+1=F(xt,ut,f(xt,ut))+wt,yt=h(xt)+vtx_{t+1} = F\bigl(x_t, u_t, f(x_t, u_t)\bigr) + w_t, \quad y_t = h(x_t) + v_t where xtx_t is the unmeasured state, utu_t the known input, f()f(\cdot) the unknown nonlinear term (parameterized by a dictionary expansion), and wt,vtw_t, v_t are process and measurement noise, respectively (Mei et al., 22 Nov 2025). The unknown dynamic function f()f(\cdot) is written as a linear combination of basis functions: f(xt,ut)=Φ(xt,ut)Tθf(x_t, u_t) = \Phi(x_t, u_t)^T \theta Here, Φ\Phi is an overcomplete library of nonlinear basis functions, and θRdθ\theta \in \mathbb{R}^{d_\theta} are the unknown weights.

AKF introduces an augmented state: xˉt=[xt θ]\bar{x}_t = \begin{bmatrix} x_t \ \theta \end{bmatrix} and a transition function Fˉ\bar{F} ensuring time-propagation of both state and model parameters: xˉt+1=Fˉ(xˉt,ut)+[wt 0]\bar{x}_{t+1} = \bar{F}(\bar{x}_t, u_t) + \begin{bmatrix} w_t \ 0 \end{bmatrix} with observations yt=h(xt)+vty_t = h(x_t) + v_t (Mei et al., 22 Nov 2025).

2. Recursive Filtering: UKF Integration and Posterior Updates

State–parameter estimation proceeds via recursive Kalman filtering in the augmented space. Prediction and update steps employ the Unscented Kalman Filter (UKF) for improved numerical stability, leveraging a joint mean–covariance propagation of both system state and model coefficients.

  • Prediction: Compute ξt\xi_t^- and Σt\Sigma_t^- for the predicted mean and covariance.
  • Update: On measurement yty_t, apply standard Kalman update with gain KtK_t:

Kt=ΣtHtT(HtΣtHtT+R)1K_t = \Sigma_t^- H_t^T (H_t \Sigma_t^- H_t^T + R)^{-1}

where HtH_t denotes the linearized measurement mapping; posterior mean and covariance are:

ξt+=ξt+Kt(yth(xt)),Σt+=(IKtHt)Σt\xi_t^+ = \xi_t^- + K_t(y_t - h(x_t^-)), \quad \Sigma_t^+ = (I - K_t H_t)\Sigma_t^-

The process is repeated over all time steps (Mei et al., 22 Nov 2025).

3. Sparse Structure Induction via Bayesian ARD

Sparsification of the model is achieved by embedding Automatic Relevance Determination (ARD) priors on the parameter block θ\theta: p(θα)=iN(θi;0,αi1)p(\theta | \alpha) = \prod_i \mathcal{N}(\theta_i; 0, \alpha_i^{-1}) Online ARD maximizes the marginal likelihood sequentially over the hyperparameters si=αi1s_i = \alpha_i^{-1} using gradient descent. The AKF posterior is updated to reflect new prior variances via a pseudo-measurement correction: qnew(xˉt)N(xˉt;ξtnew,Σtnew)q^{new}(\bar{x}_t) \sim \mathcal{N}(\bar{x}_t; \xi_t^{new}, \Sigma_t^{new}) where the correction employs standard Gaussian conditioning formulas (Mei et al., 22 Nov 2025).

This facilitates adaptive model structure selection: as ARD variances diminish, their associated basis functions are pruned, yielding an evolving, parsimonious model.

4. Algorithmic Workflow and Computational Aspects

A prototypical AKF/SKI algorithm follows these sequential steps:

  1. Measure yty_t.
  2. AKF (UKF) predict–update to obtain current posterior (ξtold,Σtold)(\xi_t^{old}, \Sigma_t^{old}).
  3. For NhpN_{hp} iterations, update prior variances via ARD gradient descent.
  4. Apply pseudo-measurement correction to posterior for new prior.
  5. Output updated mean/covariance (Mei et al., 22 Nov 2025).

Each prediction and update step incurs cubic cost in augmented state size (dx+dθ)(d_x + d_\theta) due to Cholesky factorization; ARD steps are cubic in model size dθd_\theta but remain tractable for moderate-sized dictionaries.

5. Comparative Performance and Real-World Applications

Empirical studies demonstrate AKF-based SKI methods achieve significant improvements in model selection accuracy and robustness. For instance, in WingRock benchmark experiments, mean 1\ell_1 error for SKI (AKF + ARD) reached 0.15 versus 0.95 for baseline UKF and 11.45 for basic SINDy, reflecting an 84.21% improvement (Mei et al., 22 Nov 2025). Time-delay feature selection in dynamic models is also enabled by AKF: SKI rapidly zeroes out all delays except at optimal lag, as indicated by ARD variances.

In quadrotor UAV system identification (simulated and real data), AKF with ARD consistently selected physically meaningful terms—constant, linear-PWM, and linear-drag—while standard UKF retained dense, non-interpretable coefficient blocks.

6. Limitations, Practical Guidelines, and Extensions

AKF-based SKI scales best with moderate dictionary dimensionality (dθd_\theta in tens); for larger bases, preliminary feature selection or clustering is recommended. Model non-Gaussianity may necessitate replacement of UKF with particle filters. Guidelines for tuning hyperparameters (ARD initialization, process and measurement noise covariance, step size) are provided for practical deployments.

Limitations include sensitivity to excitation richness and slow re-identification rates for highly dynamic systems. Extensions include integration of time-varying process/measurement covariances, error modeling augmentation, and use of alternative sparsification schemes.

7. Significance and Context in System Identification

AKF unifies state-observation filtering, online system identification, and interpretable Bayesian sparsification, enabling real-time, adaptive model construction under sequential data and partial measurement scenarios. The methodology obviates batch-learning requirements and “full history” data, supporting efficient state-tracking, robust parameter inference, and explicit model structure selection. Applications encompass adaptive monitoring, fault detection, and control synthesis in high-dimensional, noise-perturbed environments, with quantitative gains over baseline Kalman filtering and sparse regression (Mei et al., 22 Nov 2025).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Whiteboard

Follow Topic

Get notified by email when new papers are published related to Augmented Kalman Filter (AKF).