Finite-Arm GP-UCB
- Finite-Arm GP-UCB is a bandit algorithm that uses Gaussian process priors to balance exploration and exploitation over a fixed set of arms.
- It employs an upper confidence bound rule with computed posterior mean and variance to ensure efficient updates and strong regret guarantees.
- Enhanced variants like IRGP-UCB use randomized confidence parameters to tighten regret bounds and accelerate convergence.
Finite-Arm GP-UCB refers to Gaussian Process Upper Confidence Bound algorithms for Bayesian optimization and bandit problems where the arm set is finite. In this regime, the agent sequentially selects arms from a fixed finite set, and observations are modeled as , with following a Gaussian process (GP) prior and additive noise. Finite-armed GP-UCB provides strong regret guarantees with efficient computation, and, in recent research, variants with randomized confidence parameters have further improved regret bounds and reduced over-exploration.
1. Problem Setting: Bayesian Optimization with Finite Arms
The setting consists of a finite domain . The unknown reward function is drawn from a GP prior with mean zero and covariance satisfying for all , and the learner observes at each round , with i.i.d. noise . The aim is to minimize cumulative regret , where is an arm maximizing over (Takeno et al., 2 Sep 2024, Krishnamurthy, 20 Dec 2025).
2. The Standard Finite-Arm GP-UCB Algorithm
GP-UCB is established for maximizing an unknown function under GP priors. At each round , posterior mean and variance are computed using past data. The classic GP-UCB rule selects:
where is a deterministic confidence parameter, typically chosen to ensure a concentration inequality over all arms and all rounds, e.g. (Krishnamurthy, 20 Dec 2025, Hu et al., 11 Jun 2025). The resulting algorithm is computationally tractable on finite domains, as all required GP quantities can be efficiently updated (Hu et al., 11 Jun 2025).
3. Regret Analysis and Information Gain
A central result is that the concentration inequality
holds with a judicious choice of . Regret can be decomposed into two analytical components:
- Instance-dependent bound: Each suboptimal arm is chosen at most times, where .
- Distribution-free bound: The cumulative regret satisfies
with the maximum mutual information (information gain) between and observations, i.e.
where is the Gram matrix restricted to the set (Krishnamurthy, 20 Dec 2025, Hu et al., 11 Jun 2025). For finite , .
Table 1: Key analytical ingredients in finite-arm GP-UCB
| Ingredient | Formula/Description | Reference |
|---|---|---|
| Posterior mean | (Krishnamurthy, 20 Dec 2025) | |
| Posterior variance | (Krishnamurthy, 20 Dec 2025) | |
| Regret bound | (Krishnamurthy, 20 Dec 2025) | |
| Max info. gain | (Hu et al., 11 Jun 2025) |
4. Improved Randomized GP-UCB (IRGP-UCB)
A recognized limitation of classic GP-UCB is the need for a confidence parameter that grows with to maintain high-probability bounds. This growth is conservative and leads to excessive long-term exploration (Takeno et al., 2 Sep 2024).
IRGP-UCB replaces by a random variable sampled from a shifted exponential distribution , with and . This results in:
- The expected confidence parameter is constant in , so the width of the exploration bonus does not grow.
- The regret bound (Bayesian cumulative regret and high-probability regret) becomes
This is asymptotically sharper in than standard GP-UCB, where (Takeno et al., 2 Sep 2024). Numerical studies on synthetic and real datasets confirm IRGP-UCB achieves faster convergence and consistently lower regret.
5. Deterministic Proof Ingredients: Concentration, Radius Collapse, and Forced Deviations
The regret proof for finite-arm GP-UCB is structurally identical to that for finite-armed UCB and linear UCB algorithms (Krishnamurthy, 20 Dec 2025). The analysis consists of:
- A single uniform concentration bound (for posterior deviation),
- Radius collapse: After sufficient pulls, a suboptimal arm's posterior variance (and hence its bonus) falls below a threshold proportional to its suboptimality gap,
- Optimism-forced deviation: The selection rule cannot pick suboptimal arms after enough information is acquired, unless the confidence bound fails,
- Combining these, one derives both instance-dependent (logarithmic) and distribution-free ("oracle") regret bounds.
This structural unity extends across most optimism-based finite bandit algorithms (Krishnamurthy, 20 Dec 2025).
6. Practical Considerations and Computational Aspects
With moderate , GP-UCB and IRGP-UCB are computationally efficient. The Gram matrix can be cached, and Cholesky or inverse updates are per round (Hu et al., 11 Jun 2025). No discretization or approximate maximization is required. The regularization via the GP kernel smooths out estimates, exploiting correlations between arms for more sample-efficient exploration.
Typical applications include finite-design experimental settings, algorithm configuration, and resource allocation with categorical options (Takeno et al., 2 Sep 2024).
7. Comparison of IRGP-UCB and Classic GP-UCB
IRGP-UCB eliminates the need for time-growing confidence bonuses on finite domains. This removes late-iteration over-exploration and provides tighter bounds. In particular:
- Exploration bonus: Fixed in IRGP-UCB; grows with time in GP-UCB.
- Regret rate: for IRGP-UCB, vs. for GP-UCB.
- Empirical performance: In all reported experiments with finite , IRGP-UCB outperforms GP-UCB, Thompson Sampling, and other state-of-the-art BO algorithms in sample complexity and regret (Takeno et al., 2 Sep 2024).
A plausible implication is that, for the finite-arm setting, randomization in the confidence parameter eliminates the necessity for conservative high-probability guarantees, yielding both theoretical and practical benefits.
References: (Takeno et al., 2 Sep 2024, Hu et al., 11 Jun 2025, Krishnamurthy, 20 Dec 2025)