Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 81 tok/s
Gemini 2.5 Pro 48 tok/s Pro
GPT-5 Medium 32 tok/s Pro
GPT-5 High 32 tok/s Pro
GPT-4o 99 tok/s Pro
Kimi K2 195 tok/s Pro
GPT OSS 120B 462 tok/s Pro
Claude Sonnet 4.5 35 tok/s Pro
2000 character limit reached

Hilbert coVariance Networks (HVNs)

Updated 17 September 2025
  • Hilbert coVariance Networks (HVNs) are convolutional architectures that use the spectral properties of covariance operators in infinite-dimensional settings.
  • They employ Hilbert coVariance Filters (HVFs) through spectral and polynomial methods to extract functional principal components and higher-order relationships.
  • HVNs demonstrate robustness and transferability by effectively handling multivariate time series and functional data, outperforming traditional models in empirical validations.

Hilbert coVariance Networks (HVNs) are convolutional neural architectures constructed for signals defined over infinite-dimensional Hilbert spaces, in which processing, transformation, and representation are centered on the covariance operator rather than on pointwise kernels or finite-dimensional graph matrices. HVNs generalize covariance-based learning and convolution to settings that include functional data, multivariate time series, and reproducing kernel Hilbert spaces (RKHS), providing principled mechanisms for robust feature extraction, transferability, and the exploitation of higher-order relationships in high-dimensional or infinite-dimensional signals (Battiloro et al., 16 Sep 2025).

1. Construction of Hilbert coVariance Filters (HVFs) and HVNs

HVNs are fundamentally built from Hilbert coVariance Filters (HVFs) that transform input signals by filtering through the spectral decomposition of a covariance operator. Given a covariance operator CC on a Hilbert space HH with eigenvalues {λ}1\{\lambda_\ell\}_{\ell \geq 1} and orthonormal eigenfunctions {φ}1\{\varphi_\ell\}_{\ell \geq 1}, the canonical spectral representation is

Cv==1λv,φφvH.C v = \sum_{\ell=1}^\infty \lambda_\ell \langle v, \varphi_\ell \rangle \varphi_\ell \quad \forall v \in H.

The Hilbert coVariance Fourier Transform (HVFT) of a signal xHx \in H is the sequence of its projections onto these eigenfunctions:

x~[]=x,φ,1.\tilde{x}[\ell] = \langle x, \varphi_\ell \rangle, \quad \ell \geq 1.

A spectral HVF with frequency response h()h(\cdot) acts as

h(C)x==1h(λ)x,φφ+h(0)x,\mathfrak{h}(C) x = \sum_{\ell=1}^\infty h(\lambda_\ell) \langle x, \varphi_\ell \rangle \varphi_\ell + h(0) x_\perp,

where xx_\perp is the projection onto the kernel of CC. This operation allows pointwise manipulation of the frequency components (analogous to filtering in the frequency domain). Alternatively, a spatial (polynomial) HVF is written as

h(C)=j=0JwjCj,\mathfrak{h}(C) = \sum_{j=0}^J w_j C^j,

so that its spectral response is h(λ)=j=0Jwjλjh(\lambda) = \sum_{j=0}^J w_j \lambda^j.

HVNs are constructed by stacking such HVF banks, thereby defining layers:

xt+1u=σ(i=1Ftht(u,i)(C)xti),u=1,,Ft+1,x_{t+1}^u = \sigma\left( \sum_{i=1}^{F_t} \mathfrak{h}_t^{(u,i)}(C) x_t^i \right), \qquad u = 1, \dotsc, F_{t+1},

where σ:HH\sigma: H \rightarrow H is a pointwise (elementwise in some basis) or more generally nonlinear activation. The full HVN mapping is thus parameterized by W={HVF weights per layer and bank}\mathcal{W} = \{\text{HVF weights per layer and bank}\}, acting on CC and the initial signal collection.

The key distinction from finite-dimensional covariance neural networks (VNNs) is that the convolution/filtering operations are genuinely infinite-dimensional, relying entirely on the operator-theoretic properties of CC rather than matrix or fixed kernel constructions (Battiloro et al., 16 Sep 2025).

2. Mathematical and Operator-Theoretic Framework

The foundation of HVNs relies on properties of trace-class (or compact, self-adjoint) covariance operators, spectral integration, and discretization. For a collection of nn i.i.d. samples x1,,xnHx_1, \dotsc, x_n \in H, the empirical covariance operator is constructed as

C^nv=1ni=1nxixˉ,v(xixˉ),\hat{C}_n v = \frac{1}{n} \sum_{i=1}^n \langle x_i - \bar{x}, v \rangle (x_i - \bar{x}),

which is self-adjoint, finite-rank, and admits an eigen-decomposition with at most nn nonzero eigenvalues.

Filtering and subsequent HVN transformations are implemented using either the full spectral decomposition (if computationally feasible) or polynomial approximations, facilitating efficient computation while preserving theoretical consistency with infinite-dimensional settings. The operator filter h(C)\mathfrak{h}(C) satisfies

g~[]=h(λ)x~[],\tilde{g}[\ell] = h(\lambda_\ell) \tilde{x}[\ell],

i.e., the output in each frequency is scaled by h(λ)h(\lambda_\ell).

Importantly, HVNs guarantee the ability to exactly replicate projections onto the eigenspaces of the covariance operator using specially constructed polynomial filters. That is, for each distinct nonzero eigenvalue α\alpha of C^n\hat{C}_n, there exists a polynomial HVF hα\mathfrak{h}_\alpha such that

hα(C^n)x=Pαx,\mathfrak{h}_\alpha(\hat{C}_n) x = P_\alpha x,

where PαP_\alpha is the orthogonal projector onto the eigenspace of α\alpha. This property allows recovery of functional principal components (FPCA) by appropriately filtering the signal (Battiloro et al., 16 Sep 2025).

3. Discretization, Implementation, and Connection to Empirical Analysis

As practical implementation of HVNs requires working with finite data and computation, the framework introduces a discretization operator Sm:HRmS_m : H \rightarrow \mathbb{R}^m. For instance, in L2L^2 function spaces, SmS_m is often a bin-average map; for sequence spaces, it is projection onto the first mm coordinates; in RKHS, it might be evaluation at mm greedily chosen or equispaced points.

The empirical covariance matrix in the discrete (compressed) space is then

C^n(m)=SmC^nSm,\hat{C}_n^{(m)} = S_m \hat{C}_n S_m^*,

which ensures that empirical filtering remains consistent with the operator-level filtration, even as mm \rightarrow \infty (Proposition 1 in (Battiloro et al., 16 Sep 2025)).

The discretization allows HVNs to be deployed in a variety of settings, including but not limited to:

  • L2L^2 function spaces via binwise averaging
  • Multivariate time series
  • 2(N)\ell^2(\mathbb{N}) sequence modeling
  • Reproducing kernel Hilbert spaces, using the kernel trick and pointwise evaluations

This approach permits the transfer of mathematically rigorous infinite-dimensional filtering to practical, computationally tractable settings without sacrificing the structural connections to the underlying Hilbertian data.

4. Functional PCA Recovery and Theoretical Guarantees

HVNs possess the property that, via their spectral (or polynomial) filters, they can exactly extract the principal components of the sample covariance operator:

  • For each positive eigenvalue α\alpha of C^n\hat{C}_n, the output hα(C^n)x\mathfrak{h}_\alpha(\hat{C}_n) x is the projection of xx onto the corresponding eigenspace.
  • Taking inner products with the orthonormal eigenvectors (or functions) yields the principal scores.

This capability, proven as Theorem 1 in (Battiloro et al., 16 Sep 2025), provides strong guarantees that the network can always recover the information underlying FPCA, but can further enrich it through composition of nonlinearities and by leveraging the cross-channel covariance structure, which cannot be fully exploited by FPCA or standard multilayer perceptrons (MLPs).

The HVN framework thus generalizes classical covariance-based dimensionality reduction by embedding it in a nonlinearly parameterized, layered operator filtering and feature extraction network.

5. Robustness, Transferability, and Empirical Validation

HVNs are validated through extensive experiments on both synthetic and real-world time-series classification tasks:

  • On synthetic datasets involving multivariate Gaussian processes, HVNs outperform both MLPs and FPCA-based classifiers: the discriminative structure is captured through cross-channel covariance, which component-wise models cannot represent.
  • On real data (e.g., ECG5000), HVNs achieve consistently higher accuracy across discretization levels compared to both MLP and FPCA classifiers.

These empirical results illustrate two key benefits:

  • Robustness: HVNs are less prone to overfitting and better handle low-sample/noisy regimes due to leveraging the covariance structure in filtering.
  • Transferability: Because their architecture is parameterized via the covariance operator rather than fixed-dimensional matrices, HVNs trained under one discretization generalize to data with different resolutions or even different Hilbert space representations, reflecting versatility analogous to group-equivariant CNNs but for the setting of functional and kernel data (Battiloro et al., 16 Sep 2025).

6. Relevance to Broader Covariance-Based and Kernel Learning Paradigms

The HVN framework seamlessly unifies and extends several related lines of research:

  • Robust kernel covariance operators and cross-covariance operators as robustified building blocks for handling contaminated data (Alam et al., 2016)
  • Divergence-based comparisons between covariance operators for quantifying distributional distance between network nodes or features (Quang, 2019)
  • Operator-valued positive definite kernels and Hilbert space–valued Gaussian processes for defining layer activations and covariance-preserving features (Jorgensen et al., 23 Apr 2024)
  • Low-rank approximations and posterior covariance updates for model reduction and uncertainty quantification in infinite dimensions (Carere et al., 31 Mar 2025)
  • Stability/transferability results inherited from finite-dimensional VNNs but extended to RKHS or general Hilbert space settings (Sihag et al., 2022, Battiloro et al., 16 Sep 2025)

A plausible implication is that HVNs form a natural bridge between modern geometric deep learning, kernel methods, and (infinite-dimensional) functional data analysis. Their operator-centric design affords principled generalization, resilience to data imperfections, and efficient selection of informative subspaces for both learning and inference in high-dimensional structured domains.

7. Summary Table: Core Constructs in HVNs

Construct Mathematical Role Implementation Notes
Covariance operator CC Generates spectral decomposition, basis for filtering Empirical, trace-class
Hilbert coVariance Filter (HVF) h(C)\mathfrak{h}(C) acts spectrally or via polynomial Banked, learnable per layer
HVN layer Stacks HVFs + nonlinear activation σ\sigma Extensible to deep architectures
Discretization operator SmS_m Compresses HH to Rm\mathbb{R}^m for computation Basis- or pointwise-agnostic
FPCA recovery HVFs retrieve FPCA projections via spectral filtering Exact for sample eigenspaces
Empirical covariance matrix C^n(m)\hat{C}_n^{(m)} Implements compressed filtering Compatible with HVF design

In summary, Hilbert coVariance Networks provide a rigorously constructed, operator-theoretic extension of graph and covariance neural networks to the infinite-dimensional setting, enabling robust, transferable, and theoretically grounded learning in functional data, multivariate sequences, and RKHS. They simultaneously subsume FPCA, stabilize against overfitting, and naturally align with the mathematical structure of covariance-based statistical learning (Battiloro et al., 16 Sep 2025).

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Hilbert coVariance Networks (HVNs).