Linear Oscillatory State-Space Models
- Linear Oscillatory State-Space Models (LinOSS) are structured state-space models using forced harmonic oscillators to provide an analytically tractable framework for Gaussian process modeling.
- They unify Matérn and damped SHO processes via continuous and discrete formulations, enabling efficient long-range sequence learning with precise hyperparameter control.
- Variants like LinOSS-IM, IMEX, and D-LinOSS demonstrate competitive performance on tasks such as time-series classification and regression, leveraging fast associative scans for reduced computational complexity.
Linear Oscillatory State-Space Models (LinOSS) represent a class of structured state-space models formed by discrete layers of forced harmonic oscillators. They provide an analytically tractable framework unifying Matérn and damped simple harmonic oscillator (SHO) Gaussian processes under a linear, Gaussian, Markovian state-space representation (Jordán et al., 2021). Recent neural architectures extend LinOSS principles to the efficient modeling of extremely long-range dependencies and to high-capacity sequence learning (Rusch et al., 4 Oct 2024, Boyer et al., 17 May 2025).
1. Continuous-Time Formulation
LinOSS models encode latent Markov processes driven by white noise, observed via a linear projection . In differential notation,
where is the state matrix, couples a multidimensional Wiener process , and selects observed components. For the SHO, with natural frequency and damping factor , the dynamics are second-order with state dimension ,
This construction generalizes to Matérn covariances for smoothness parameters , where state dimension and system matrices follow explicit block constructions for each (Jordán et al., 2021). In all cases, LinOSS provide rational spectral densities with a closed-form correspondence between state-space parameters and kernel hyperparameters. In modern neural sequence models, LinOSS blocks are built from networks of uncoupled forced oscillators (Rusch et al., 4 Oct 2024, Boyer et al., 17 May 2025):
where is diagonal, is the hidden state, and is the input-to-state coupling.
2. Discretization and Algorithmic Structure
Discrete-time LinOSS recurrences are obtained via exact matrix exponentiation for classic Gaussian process applications (Jordán et al., 2021):
with defined by an integral Lyapunov equation. For neural networks, structure-preserving implicit-explicit (IMEX) and fully implicit (IM) Euler schemes are applied (Rusch et al., 4 Oct 2024, Boyer et al., 17 May 2025):
IM scheme: , yielding a linear recurrence
IMEX scheme: updates preserve symplecticity (volume) and time-reversibility with
and block-wise recurrence .
Efficient sequence processing is realized via fast associative parallel scans (tree-scan algorithms), which reduce wall-clock depth from to and allow GPU kernels to scale to sequences of length (Rusch et al., 4 Oct 2024).
3. Covariance, Spectral Structure, and Hyperparameterization
Analytically, LinOSS latent covariances solve the Lyapunov equation,
the output covariance function is
recovering the Matérn and SHO kernel forms (Jordán et al., 2021).
The spectral density for LinOSS is obtained by Fourier transform:
with SHO poles at . For Matérn , the density is
with repeated poles at .
Hyperparameters for LinOSS models include amplitude , characteristic time scale or , and damping/smoothness or . For neural architectures, frequency parameters are constrained for stability via nonnegativity and may be initialized via uniform sampling, regularized, and learned as free variables (Rusch et al., 4 Oct 2024).
4. Universality, Stability, and Extensions
LinOSS blocks are universal function approximators for causal, continuous operators between time-varying inputs and outputs, provably capable of arbitrarily accurate approximation via suitably chosen oscillator frequencies and readout mappings (Rusch et al., 4 Oct 2024). Nonnegativity of oscillator frequencies ensures global asymptotic stability in IM schemes, while IMEX schemes retain symplecticity and invertibility—allowing memory-efficient backpropagation.
However, canonical LinOSS variants (IM, IMEX) couple damping and frequency rigidly: only one dissipation timescale per oscillator, limiting representational flexibility for certain dynamics. To address this, Damped LinOSS (D-LinOSS) models allow independent, learnable per-mode damping parameters (Boyer et al., 17 May 2025). In D-LinOSS, damping and frequency are diagonal and unconstrained within the stability region , allowing the full complex unit disk of eigenvalues to be realized. This decoupling enables superior representation of multi-timescale and long-range dependencies.
5. Computational Complexity and Memory
Sequential LinOSS models scale with (for oscillator dimension , sequence length ). With parallel associative scan algorithms, processing depth is reduced to at total flops (Rusch et al., 4 Oct 2024). For Gaussian process inference, likelihood evaluation and Kalman filtering are , or for sparse/block-diagonal (Jordán et al., 2021).
Memory usage benefits from invertibility in IMEX: only checkpointed activations are required for backpropagation, reducing from to plus log-sized checkpoints. IM variants require standard checkpointing or backscan reversal.
6. Empirical Performance and Model Variants
Empirical results across multiple benchmarks demonstrate LinOSS’s competitive and sometimes superior performance versus state-of-the-art sequence models:
- UEA MTS Classification: LinOSS-IM achieves 67.8% mean test accuracy (max 95% on 17,984-length sequence), beating Log-NCDE and S5 (Rusch et al., 4 Oct 2024).
- PPG-DaLiA Regression: LinOSS-IM MSE $0.064$ (), outperforming LRU ($0.156$), S5 ($0.128$), Mamba ($0.107$) (Rusch et al., 4 Oct 2024).
- Long-Horizon Weather Forecast: LinOSS-IMEX MAE $0.508$, LinOSS-IM $0.528$, S4 $0.578$, outperforming transformer models (Rusch et al., 4 Oct 2024).
- D-LinOSS Comparisons: D-LinOSS achieves improved RMSE ( vs for LinOSS-IM), higher accuracy, and reduced hyperparameter search space, due to decoupled damping (Boyer et al., 17 May 2025).
Architectural variants include LinOSS-IM (dissipative, forgetting), LinOSS-IMEX (volume-preserving), and D-LinOSS (learnable, multi-timescale dissipation). Training employs GLU regularization, residual connections, and time-indexed input. Integration of time-varying layers and domain-specific oscillatory priors remains an open route for further paper.
7. Connections to Gaussian Process Theory
LinOSS models are grounded in state-space representations of Gaussian processes with rational spectral densities, providing closed-form kernels and efficient inference via standard Kalman algorithms. The framework allows seamless unification of Matérn and damped oscillator processes, continuous and discrete-time models, and supports fast hyperparameter learning by maximizing likelihood or MCMC (Jordán et al., 2021).
The construction elucidates the spectral–temporal relationship of kernel smoothness, impulse response, and damping, and provides closed-form expressions for covariance and spectral densities relevant to time series modeling, astronomical data analysis, and general sequence learning (Jordán et al., 2021). The extension to neural architectures preserves computational efficiency while unlocking new representational capacities essential for modeling long-sequence dynamics.