Hilbert coVariance Networks (HVNs)
- Hilbert coVariance Networks (HVNs) are convolutional architectures that use the spectral properties of covariance operators in infinite-dimensional settings.
- They employ Hilbert coVariance Filters (HVFs) through spectral and polynomial methods to extract functional principal components and higher-order relationships.
- HVNs demonstrate robustness and transferability by effectively handling multivariate time series and functional data, outperforming traditional models in empirical validations.
Hilbert coVariance Networks (HVNs) are convolutional neural architectures constructed for signals defined over infinite-dimensional Hilbert spaces, in which processing, transformation, and representation are centered on the covariance operator rather than on pointwise kernels or finite-dimensional graph matrices. HVNs generalize covariance-based learning and convolution to settings that include functional data, multivariate time series, and reproducing kernel Hilbert spaces (RKHS), providing principled mechanisms for robust feature extraction, transferability, and the exploitation of higher-order relationships in high-dimensional or infinite-dimensional signals (Battiloro et al., 16 Sep 2025).
1. Construction of Hilbert coVariance Filters (HVFs) and HVNs
HVNs are fundamentally built from Hilbert coVariance Filters (HVFs) that transform input signals by filtering through the spectral decomposition of a covariance operator. Given a covariance operator on a Hilbert space with eigenvalues and orthonormal eigenfunctions , the canonical spectral representation is
The Hilbert coVariance Fourier Transform (HVFT) of a signal is the sequence of its projections onto these eigenfunctions:
A spectral HVF with frequency response acts as
where is the projection onto the kernel of . This operation allows pointwise manipulation of the frequency components (analogous to filtering in the frequency domain). Alternatively, a spatial (polynomial) HVF is written as
so that its spectral response is .
HVNs are constructed by stacking such HVF banks, thereby defining layers:
where is a pointwise (elementwise in some basis) or more generally nonlinear activation. The full HVN mapping is thus parameterized by , acting on and the initial signal collection.
The key distinction from finite-dimensional covariance neural networks (VNNs) is that the convolution/filtering operations are genuinely infinite-dimensional, relying entirely on the operator-theoretic properties of rather than matrix or fixed kernel constructions (Battiloro et al., 16 Sep 2025).
2. Mathematical and Operator-Theoretic Framework
The foundation of HVNs relies on properties of trace-class (or compact, self-adjoint) covariance operators, spectral integration, and discretization. For a collection of i.i.d. samples , the empirical covariance operator is constructed as
which is self-adjoint, finite-rank, and admits an eigen-decomposition with at most nonzero eigenvalues.
Filtering and subsequent HVN transformations are implemented using either the full spectral decomposition (if computationally feasible) or polynomial approximations, facilitating efficient computation while preserving theoretical consistency with infinite-dimensional settings. The operator filter satisfies
i.e., the output in each frequency is scaled by .
Importantly, HVNs guarantee the ability to exactly replicate projections onto the eigenspaces of the covariance operator using specially constructed polynomial filters. That is, for each distinct nonzero eigenvalue of , there exists a polynomial HVF such that
where is the orthogonal projector onto the eigenspace of . This property allows recovery of functional principal components (FPCA) by appropriately filtering the signal (Battiloro et al., 16 Sep 2025).
3. Discretization, Implementation, and Connection to Empirical Analysis
As practical implementation of HVNs requires working with finite data and computation, the framework introduces a discretization operator . For instance, in function spaces, is often a bin-average map; for sequence spaces, it is projection onto the first coordinates; in RKHS, it might be evaluation at greedily chosen or equispaced points.
The empirical covariance matrix in the discrete (compressed) space is then
which ensures that empirical filtering remains consistent with the operator-level filtration, even as (Proposition 1 in (Battiloro et al., 16 Sep 2025)).
The discretization allows HVNs to be deployed in a variety of settings, including but not limited to:
- function spaces via binwise averaging
- Multivariate time series
- sequence modeling
- Reproducing kernel Hilbert spaces, using the kernel trick and pointwise evaluations
This approach permits the transfer of mathematically rigorous infinite-dimensional filtering to practical, computationally tractable settings without sacrificing the structural connections to the underlying Hilbertian data.
4. Functional PCA Recovery and Theoretical Guarantees
HVNs possess the property that, via their spectral (or polynomial) filters, they can exactly extract the principal components of the sample covariance operator:
- For each positive eigenvalue of , the output is the projection of onto the corresponding eigenspace.
- Taking inner products with the orthonormal eigenvectors (or functions) yields the principal scores.
This capability, proven as Theorem 1 in (Battiloro et al., 16 Sep 2025), provides strong guarantees that the network can always recover the information underlying FPCA, but can further enrich it through composition of nonlinearities and by leveraging the cross-channel covariance structure, which cannot be fully exploited by FPCA or standard multilayer perceptrons (MLPs).
The HVN framework thus generalizes classical covariance-based dimensionality reduction by embedding it in a nonlinearly parameterized, layered operator filtering and feature extraction network.
5. Robustness, Transferability, and Empirical Validation
HVNs are validated through extensive experiments on both synthetic and real-world time-series classification tasks:
- On synthetic datasets involving multivariate Gaussian processes, HVNs outperform both MLPs and FPCA-based classifiers: the discriminative structure is captured through cross-channel covariance, which component-wise models cannot represent.
- On real data (e.g., ECG5000), HVNs achieve consistently higher accuracy across discretization levels compared to both MLP and FPCA classifiers.
These empirical results illustrate two key benefits:
- Robustness: HVNs are less prone to overfitting and better handle low-sample/noisy regimes due to leveraging the covariance structure in filtering.
- Transferability: Because their architecture is parameterized via the covariance operator rather than fixed-dimensional matrices, HVNs trained under one discretization generalize to data with different resolutions or even different Hilbert space representations, reflecting versatility analogous to group-equivariant CNNs but for the setting of functional and kernel data (Battiloro et al., 16 Sep 2025).
6. Relevance to Broader Covariance-Based and Kernel Learning Paradigms
The HVN framework seamlessly unifies and extends several related lines of research:
- Robust kernel covariance operators and cross-covariance operators as robustified building blocks for handling contaminated data (Alam et al., 2016)
- Divergence-based comparisons between covariance operators for quantifying distributional distance between network nodes or features (Quang, 2019)
- Operator-valued positive definite kernels and Hilbert space–valued Gaussian processes for defining layer activations and covariance-preserving features (Jorgensen et al., 23 Apr 2024)
- Low-rank approximations and posterior covariance updates for model reduction and uncertainty quantification in infinite dimensions (Carere et al., 31 Mar 2025)
- Stability/transferability results inherited from finite-dimensional VNNs but extended to RKHS or general Hilbert space settings (Sihag et al., 2022, Battiloro et al., 16 Sep 2025)
A plausible implication is that HVNs form a natural bridge between modern geometric deep learning, kernel methods, and (infinite-dimensional) functional data analysis. Their operator-centric design affords principled generalization, resilience to data imperfections, and efficient selection of informative subspaces for both learning and inference in high-dimensional structured domains.
7. Summary Table: Core Constructs in HVNs
Construct | Mathematical Role | Implementation Notes |
---|---|---|
Covariance operator | Generates spectral decomposition, basis for filtering | Empirical, trace-class |
Hilbert coVariance Filter (HVF) | acts spectrally or via polynomial | Banked, learnable per layer |
HVN layer | Stacks HVFs + nonlinear activation | Extensible to deep architectures |
Discretization operator | Compresses to for computation | Basis- or pointwise-agnostic |
FPCA recovery | HVFs retrieve FPCA projections via spectral filtering | Exact for sample eigenspaces |
Empirical covariance matrix | Implements compressed filtering | Compatible with HVF design |
In summary, Hilbert coVariance Networks provide a rigorously constructed, operator-theoretic extension of graph and covariance neural networks to the infinite-dimensional setting, enabling robust, transferable, and theoretically grounded learning in functional data, multivariate sequences, and RKHS. They simultaneously subsume FPCA, stabilize against overfitting, and naturally align with the mathematical structure of covariance-based statistical learning (Battiloro et al., 16 Sep 2025).