Functional Gradient Ascent (FGA)
- Functional Gradient Ascent (FGA) is a method for optimizing infinite-dimensional function inputs by computing Fréchet derivatives on acquisition functions derived from Gaussian process models.
- It employs a scalarized upper confidence bound (UCB) acquisition approach with integrated weight functions to effectively navigate high-dimensional function spaces.
- Empirical benchmarks demonstrate that FGA significantly reduces regret and improves sample efficiency compared to alternative optimization methods in scientific and engineering applications.
The functional gradient ascent algorithm (FGA) is a method for optimizing infinite-dimensional function-valued inputs over acquisition functions derived from a function-on-function Gaussian process (FFGP) model within function-on-function Bayesian optimization (FFBO). This setting arises when both the input and output of the optimization are elements of functional spaces, frequently encountered in advanced scientific and engineering domains requiring optimization of curves, shapes, or other high-dimensional function-valued objects (Huang et al., 16 Nov 2025).
1. Mathematical Formulation of Function-on-Function BO
In FFBO, the objective is to maximize a black-box map , where is an infinite-dimensional Hilbert space of square-integrable functions defined on and is the output space (also a Hilbert space of functions on ). The principal challenge lies in modeling posterior uncertainty and optimizing over .
The FFGP prior is defined by a mean and a separable operator-valued kernel
where is a positive-definite scalar-valued kernel (e.g., Matérn kernel on the metric), and is a nonnegative self-adjoint operator on (typically an integral operator).
For observation pairs , the posterior mean and covariance functions of are given by
where is an block-matrix of operator-valued kernels and is the functional output noise variance.
2. Scalarized Acquisition and UCB Functional
A scalar acquisition function is defined through a functional mapping to a scalar by
for a weight function . The scalarized mean and variance become
where and is the kernel underlying .
The upper confidence bound (UCB) acquisition function is then
with exploration parameter .
3. Functional Gradient Ascent Algorithm
FGA is applied to maximize over the infinite-dimensional input space: The Fréchet derivative (the functional generalization of the gradient) is computed as: where
For the -distance based kernel (e.g., Matérn), the Fréchet derivative with respect to variations is
The optimization is performed as an iterative ascent, i.e.,
for step-size until convergence.
4. Theoretical Properties and Convergence
Under regularity conditions (Matérn kernel, trace-class operator , bounded norms, and appropriate vanishing noise scaling), the posterior for remains well-defined, and truncation at finite rank yields
(regardless of the choice of basis). The FFBO regret with FGA satisfies, with high probability,
with information gain , thus simple regret is . This provides a nontrivial guarantee for global optimization in infinite-dimensional settings (Huang et al., 16 Nov 2025).
5. Empirical Performance and Benchmarks
FGA-based FFBO is evaluated against baselines: FIBO (function-input BO), FOBO (function-output BO), and MTBO (multi-task BO). On both synthetic and real-world tasks involving optimization over function-valued domains (e.g., one-dimensional curves, stress–strain waveform matching), FFBO achieves faster convergence and lower regret across all tested scenarios for the same query budget, with robustness to noise and consistent superiority in sample efficiency (Huang et al., 16 Nov 2025).
6. Context within Operator-Based Bayesian Optimization
Alternative approaches in the operator learning and surrogate modeling literature include surrogate construction via parametric operator networks (e.g., NEON, as in (Guilhoto et al., 3 Apr 2024)), which operate over deterministic mappings and use backpropagation-based optimizers (e.g., L-BFGS) in the reduced design space . These methods contrast with FGA in their reliance on finite-parametric representations and backpropagation instead of Fréchet-based optimization over function spaces.
7. Implications and Extensions
FGA represents a principled approach for functional optimization in the fully infinite-dimensional regime, circumventing intrinsic limitations of finite-parameterization. A plausible implication is that FGA enables direct exploitation of smoothness and structural priors on function spaces—reflected in the kernel and operator choices—yielding both theoretical and empirical superiority in problems where function-valued inputs and outputs are intrinsic (Huang et al., 16 Nov 2025). Such methods set the foundation for further advances in functional sequential design, shape optimization, and scientific computing with high-dimensional function spaces.
Sponsored by Paperpile, the PDF & BibTeX manager trusted by top AI labs.
Get 30 days free