Papers
Topics
Authors
Recent
2000 character limit reached

Infinite-Dimensional Gaussian Processes

Updated 24 December 2025
  • Infinite-dimensional Gaussian processes are stochastic models defined on function spaces with Gaussian measures and trace-class covariance operators.
  • They bridge advanced frameworks, linking RKHS, optimal transport geometry, and functional regression to support scalable uncertainty quantification.
  • Applications include operator learning in stochastic PDEs, Bayesian inversion, and adaptive Monte Carlo methods for high-dimensional inference.

An infinite-dimensional Gaussian process is a probabilistic object whose sample paths, parameterized by time, space, or more general indices, take values in infinite-dimensional function spaces such as separable Hilbert or Banach spaces. Such processes are characterized by laws that are Gaussian measures on these topological vector spaces, equipped with covariance operators of trace-class. Infinite-dimensional Gaussian processes are foundational in stochastic analysis, statistical learning theory, Bayesian field inference, operator learning, and high-dimensional uncertainty quantification. The mathematical and algorithmic infrastructure for their study includes measure-theoretic constructions, operator theory, stochastic PDEs, infinite-dimensional statistical geometry, optimal transport, and numerical methods for function-aligned inference.

1. Infinite-Dimensional Gaussian Measure Framework

An infinite-dimensional Gaussian process is defined as a collection {Xt}tD\{X_t\}_{t \in D} (or, more generally, random elements XX in a separable Hilbert space HH) such that every finite linear combination of the coordinates is (real) Gaussian N(0,σ2)N(0,\sigma^2). The law of XX can be identified with a Borel Gaussian measure μP2(H)\mu \in \mathcal{P}_2(H), specified by a mean mHm \in H and a positive, self-adjoint, trace-class covariance operator C:HHC : H \to H (Quang, 2020, Wallin et al., 2018). For 1,2H\ell_1,\ell_2 \in H^*, the covariance is

u,Cv=Hxm,uxm,vdμ(x)\langle u, C v \rangle = \int_H \langle x-m, u \rangle \langle x-m, v \rangle \, d\mu(x)

and induces a Cameron–Martin space E=Range(C1/2)E = \text{Range}(C^{1/2}) with inner product u,vE=C1/2u,C1/2v\langle u, v \rangle_E = \langle C^{-1/2}u, C^{-1/2}v \rangle. The Karhunen–Loève expansion is

X=m+k=1λkξkekX = m + \sum_{k=1}^\infty \sqrt{\lambda_k}\, \xi_k\, e_k

with λk,ek\lambda_k, e_k the eigenpairs of CC and ξkN(0,1)\xi_k \sim N(0,1) independent (Wallin et al., 2018). Establishing this infinite-dimensional foundation is essential for advanced modeling of function-valued phenomena, nonparametric Bayes, adaptive Monte Carlo algorithms, and operator inference.

2. Covariance Structure and RKHS Connections

The covariance kernel K(x,y)=E[XxXy]K(x, y) = \mathbb{E}[X_x X_y] associated to a Gaussian process induces a reproducing kernel Hilbert space (RKHS) HK\mathcal{H}_K on the index set DD (Jorgensen et al., 2022). Given any positive-definite kernel K:D×DRK : D \times D \to \mathbb{R}, there exists a unique, centered Gaussian process XX whose L2L^2 closure of finite linear combinations of XtX_{t} is isomorphic to HK\mathcal{H}_K (Jorgensen et al., 2022). The Cameron–Martin space of the Gaussian measure coincides with the RKHS, and point evaluation is continuous. Isomorphism theorems exist between measure spaces of finite (signed) energy and HK\mathcal{H}_K via TKμ=K(s,)dμ(s)T_K \mu = \int K(s, \cdot)\, d\mu(s), and between L2L^2 path spaces and HK\mathcal{H}_K (infinite-dimensional Fourier transform) via TF(t)=E[eiXtF]\mathcal{T} F(t) = \mathbb{E}[e^{-i X_t} F] (Jorgensen et al., 2022), enabling spectral and pathwise analysis on stochastic function spaces.

3. Conditioning, Disintegration, and Infinite-Dimensional Regression

Given a Gaussian measure μ\mu on a Banach space XX (e.g., C(D)C(D)), linear observations y=G(x)y = G(x) for bounded operators G:XYG : X \to Y admit an explicit conditional Gaussian measure, or disintegration, provided the pushforward covariance Cν=GCμGC_\nu = G C_\mu G^* is of finite rank (2207.13581, LaGatta, 2010). The conditional mean is affine in yy, and the conditional covariance is independent of yy. This yields consistent update rules for Gaussian process regression even when the observation operators are infinite-dimensional or involve linear functionals, as in functional data analysis or PDE-constrained Bayesian inversion: m1=m0+C0G(GC0G)1(yGm0),C1=C0C0G(GC0G)1GC0m_1 = m_0 + C_0\, G^* (G\, C_0\, G^*)^{-1}(y - G m_0), \quad C_1 = C_0 - C_0\, G^* (G\, C_0\, G^*)^{-1} G C_0 (2207.13581). These equations generalize the Kalman filter and Bayesian posterior update to Hilbert spaces. Continuous disintegration is ensured precisely when a certain operator norm MM is finite (LaGatta, 2010).

4. Entropic Regularization and Optimal Transport Geometry

Entropic optimal transport on infinite-dimensional Hilbert spaces provides a differentiable interpolation between squared L2L^2-distance and classical quadratic Wasserstein distance, addressing the singularities and ill-posedness of the exact W2W_2 metric in infinite dimensions (Quang, 2020). For Gaussian measures μ0=N(m0,C0)\mu_0 = N(m_0, C_0) and μ1=N(m1,C1)\mu_1 = N(m_1, C_1), the entropic-regularized cost is

OT2ϵ(μ0,μ1)=infγEγxy2+ϵKL(γμ0μ1)\mathrm{OT}_{2}^\epsilon (\mu_0, \mu_1) = \inf_{\gamma} \mathbb{E}_\gamma \|x-y\|^2 + \epsilon \, \mathrm{KL}(\gamma \| \mu_0 \otimes \mu_1)

whose unique minimizer is a joint Gaussian coupling with explicit cross-covariance, and a closed-form formula for the cost in terms of the means and covariances. The associated Sinkhorn divergence S2ϵS_2^\epsilon is strictly positive definite and Fréchet differentiable everywhere, in contrast to W2W_2 (Quang, 2020).

For sets of Gaussian measures (e.g., for Bayesian model averaging or dictionary learning in function space), the Sinkhorn barycenter has a unique solution in Hilbert spaces, given by a matrix fixed-point equation involving the input covariances (Quang, 2020). In the special case of Gaussian measures on RKHSs, the entropic-regularized OT and Sinkhorn divergences admit explicit kernel matrix expressions, interpolating between the kernel MMD and the Wasserstein metric.

5. Construction and Analysis of Kernels in Distribution Spaces

To index GPs on probability distributions, optimal transport-based Hilbertian embeddings have been developed (Bachoc et al., 2018). Measures μ\mu are mapped to L2(μˉ)L^2(\bar{\mu}) via optimal transport to a Wasserstein barycenter, linearizing the Wasserstein geometry for kernel construction. Radial kernels F(Φ(μ)Φ(ν))F(\|\Phi(\mu) - \Phi(\nu)\|) where FF is completely monotone generate positive definite kernels on L2L^2-spaces, ensuring strict positive definiteness, rotational invariance, and microergodicity (every distinct parameter induces singular laws in infinite dimensions). Consistency holds for empirical barycenters and such approaches enable the extension of regression techniques to spaces of distributions (Bachoc et al., 2018).

6. Operator-Valued and Pathwise Constructions

Infinite-dimensional Gaussian processes arise as solutions of infinite-dimensional stochastic differential equations (SDEs), including SPDEs and Volterra-type integrals driven by Hilbert-valued Gaussian noise (Benth et al., 2020, Beghin et al., 2023). The covariance of such pathwise constructions can be analyzed using 2D Volterra sewing lemmas, yielding operator-valued covariance functionals. Fractional, Ornstein–Uhlenbeck, and rough path extensions are naturally accommodated. Generalized fractional operators parameterized by Bernstein functions permit modeling of long- and short-range dependence, persistence, and local time properties (Beghin et al., 2023). Regularity, entropy, and the spectrum of these covariances are central in both the existence of rough-path lifts and the design of MCMC algorithms for functional Bayesian inference (Wallin et al., 2018).

7. Applications, Computation, and Inductive Bias

Infinite-dimensional Gaussian processes underpin kernel methods for functional data, stochastic PDE modeling, operator learning, and uncertainty quantification in distributed systems (Souza et al., 19 Oct 2025, Courteville et al., 15 Dec 2025). Implementations crucially exploit discretized representers from the underlying Hilbert or RKHS structure, adaptive MCMC in function spaces, and the differentiability properties of regularized transport for high-dimensional optimization. Neural operator limits, tensor network GP analogues, and port-Hamiltonian system surrogates demonstrate the breadth of inductive biases and architectures that can be embedded within principled infinite-dimensional GP frameworks (Souza et al., 19 Oct 2025, Guo et al., 2021, Courteville et al., 15 Dec 2025).

In summary, infinite-dimensional Gaussian processes unify advanced probabilistic modeling on function spaces, computational methods in high- or infinite-dimensional regimes, regularized transport geometry, and functional operator learning, with substantial theoretical infrastructure supporting practical, scalable, and well-posed inference (Quang, 2020, Bachoc et al., 2018, Beghin et al., 2023, Benth et al., 2020, Jorgensen et al., 2022, Souza et al., 19 Oct 2025, Courteville et al., 15 Dec 2025).

Whiteboard

Follow Topic

Get notified by email when new papers are published related to Infinite-Dimensional Gaussian Processes.