Upper Confidence Bounds with Curiosity (UCC)
- Upper Confidence Bounds with Curiosity (UCC) is a framework that augments classic UCB algorithms with dynamic, curiosity-driven exploration to balance exploitation and adaptive exploration.
- It synthesizes theory and practice by integrating information-theoretic measures and rational models to adjust exploration bonuses based on uncertainty and learning progress.
- UCC methods improve sample efficiency and robustness in complex settings such as stochastic bandits, reinforcement learning, and heavy-tailed environments.
Upper Confidence Bounds with Curiosity (UCC) designates a class of exploration strategies for sequential decision making—most notably in stochastic multi-armed bandits and reinforcement learning—where classical Upper Confidence Bound (UCB) algorithms are augmented by mechanisms that instantiate curiosity-driven or adaptive exploration. UCC-type methods synthesize the classic optimism-in-the-face-of-uncertainty principle with more dynamic or information-theoretically inspired bonuses drawn from empirical or theoretical measures of uncertainty, information gain, or learning progress. The resulting framework encompasses a wide range of instantiations in bandit problems, contextual bandits, reinforcement learning, and more general stochastic process optimization. Approaches to UCC are informed by both lower bound theory and rational models of curiosity.
1. Mathematical Structure and Foundational UCB Forms
The archetypal UCB policy for stochastic bandits assigns, at each round , to arm an index
where is the empirical mean after pulls, and is a tunable constant. This index achieves a high-probability upper bound for the unknown mean reward and balances exploitation (via ) with exploration (via ). Extensions such as UCB replace the constant by a parameter :
The parameter modulates the amount of forced exploration: larger increases exploration, crucial for worst-case optimality. Generalized policies consider for any increasing function , capturing a trade-off between sample efficiency in "hard" versus "simple" environments (Salomon et al., 2011).
In KL-UCB algorithms, the bonus is based on inverting the Kullback-Leibler divergence between the empirical distribution and plausible means ,
where is the appropriate divergence for the underlying parametric family (Cappé et al., 2012). This tightens confidence intervals and aligns the bonus with the statistical geometry of the model.
When ported to reinforcement learning, especially in high-dimensional or continuous spaces, the UCB principle is operationalized through model ensembles, random feature embeddings (e.g., via neural tangent kernel gradients), or by optimizing with respect to uncertainty in latent model predictions (Chen et al., 2017, Zhou et al., 2019, Seyde et al., 2020).
2. Regret Lower Bounds and Adaptive Exploration
UCB methods are characterized by their regret scaling properties. Consistent policies, as defined by Lai and Robbins, guarantee
and incur a minimal number of explorative pulls per suboptimal arm, set by the instance-dependent complexity ,
where is a min-KL subject to mean constraints. For generalized policies (so-called -consistent), the bound is relaxed to
highlighting the exploration-sample complexity trade-off (Salomon et al., 2011).
Crucially, more aggressive adaptations—in which grows sub-logarithmically—may minimize regret in "easy" (e.g., deterministic) environments down to , but at the cost of potentially catastrophic regret in "hard" settings. Theoretical impossibility results preclude the existence of adaptive meta-algorithms that optimally select the best exploration rate online without environmental side information (Salomon et al., 2011).
3. Curiosity-Driven Bonuses: Rational Models and Information-Theoretic Extensions
Curiosity in UCC extends exploration strategies by dynamically tuning the exploration term using either empirical uncertainty, theoretical information gain estimates, or intrinsic learning progress. Rational models (Dubey et al., 2017) formalize curiosity as the expected change in knowledge utility ,
where is the probability of need ("need probability"), the exposure, and the confidence in stimulus . Mathematical forms such as (inverted-U w.r.t. confidence) or (novelty-driven) provide dynamic curiosity bonuses as a function of confidence and sampling history.
Mechanistically, this bonus can be integrated into classical UCB by augmenting the confidence bound, yielding policies of the form:
where quantifies local curiosity and balances exploitation, uncertainty, and exploration according to context.
In information-theoretic UCB extensions, such as KL-UCB, the width of the confidence region already quantifies the "difficulty" of the learning problem through divergence. This approach can be merged with curiosity by, for instance, adaptively increasing the exploration parameter for arms or states where learning progress is maximal or statistical surprise is high (Cappé et al., 2012).
4. UCC in Reinforcement Learning and Contextual Bandits
In deep RL, UCC is operationalized through ensemble methods and context-dependent variance estimation:
- Q-ensemble approaches maintain an ensemble of independently learned -functions , computing
and selecting actions via
which encodes optimism and directs exploration to uncertain state-action pairs (Chen et al., 2017).
- In context-dependent UCBs for function approximation (e.g., UCLS), the local covariance of value estimates is tracked, yielding state-action specific confidence bounds and focusing exploration where uncertainty or prediction error is large (Kumaraswamy et al., 2018).
- NeuralUCB leverages neural tangent kernel embeddings: for each context , define , and build UCBs of the form
thus quantifying uncertainty as a function of learned distributed representations (Zhou et al., 2019).
- In contextual bandits with very large or infinite action spaces, UCCB (Upper Confidence Counterfactual Bounds) constructs confidence intervals in policy space rather than just in action space, leveraging counterfactual divergence and context complexity to achieve optimal trade-offs without dependency on context cardinality (Xu et al., 2020).
5. Implications of Prior Information, Optimality, and Limiting Theorems
Bayesian perspectives on UCC (e.g., (Russo, 2019, Atsidakou et al., 2023)) show that Upper Confidence Bound rules can be formally equivalent to Gittins indices in certain regimes, especially in Gaussian bandits with large time horizons (discount factor ):
This establishes that proper tuning of the optimism parameter (via the linkage for the quantile) can achieve the same exploration-exploitation characteristics as dynamically optimal solutions.
Recent results further establish that, under certain conditions, UCB-type procedures with exploration bonuses tuned by prior informativeness (in terms of variance and mean separation) can achieve finite-time regret that is logarithmic or even constant in favorable regimes,
with explicit dependency on gap surprise and the prior's information content (Atsidakou et al., 2023).
The impossibility of oracle selectivity imposes that UCC designs must trade-off between uniform regret guarantees and specialization to narrow classes of environments—no algorithm can "meta-optimize" exploration adaptively to the unknown environment class (Salomon et al., 2011).
6. Robustness, Heavy-Tailed Distributions, and Data-Driven UCB
Standard UCBs rely on moment assumptions (variance, sub-Gaussianity) that may not hold in practical, heavy-tailed environments. Data-driven UCBs using robust estimators such as the resampled median-of-means (RMM) circumvent this limitation by constructing parameter-free, distribution-free confidence bounds without recourse to unknown moment parameters (Tamás et al., 9 Jun 2024). The procedure for RMM-UCB constructs bounds directly from empirical data using randomized sign flips and median aggregation, supporting robust operation even when rewards are heavy-tailed, thereby enabling curiosity-based bonuses to be reliably computed in non-sub-Gaussian regimes.
The associated regret bounds (for symmetric heavy-tailed settings) remain near optimal:
independent of unknown distributional moments, providing strong guarantees in adversarial noise.
7. Practical Deployment and Application Scenarios
Applications of UCC-type algorithms are diverse:
- In global optimization of continuous domains, the use of generic chaining UCBs (with discretization trees) extends optimism- and curiosity-driven methods to Gaussian or broader stochastic processes (Contal et al., 2016). These approaches adapt exploration granularity to intrinsic problem complexity (e.g., the metric entropy of the domain).
- In robot planning or point-goal navigation, model uncertainty from an ensemble of predictors is integrated via UCB objectives, selecting actions or paths that are both promising and uncertain—an instantiation of curiosity at the planning level (Georgakis et al., 2022). For example, path costs may be minimized as
where is mean traversability, is epistemic uncertainty, and is path length.
- In model-based RL, latent ensemble models forecast long-term returns, and a UCB objective prioritizes both high-mean and high-variance (curious) exploratory actions, yielding substantial improvements in sample efficiency and robust learning (Seyde et al., 2020).
Conclusion
UCC, as an umbrella for Upper Confidence Bounds with Curiosity, encapsulates the progressive refinement of optimism-based exploration methods through principled and dynamic uncertainty quantification, adaptation to structural properties of the environment, and robust, data-driven estimation. While universal adaptivity is theoretically precluded, approaches integrating information-theoretic, rational, or empirical curiosity can sharply improve sample efficiency and robustness within target environment classes. Leading algorithmic instantiations span bandits, contextual bandits with deep representations, reinforcement learning with ensembles, robust statistics, and planning under epistemic uncertainty, each demonstrating how curiosity—mathematically formalized—augments the power and nuance of exploration in learning agents.