Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Online learning with kernel losses (1802.09732v1)

Published 27 Feb 2018 in stat.ML and cs.LG

Abstract: We present a generalization of the adversarial linear bandits framework, where the underlying losses are kernel functions (with an associated reproducing kernel Hilbert space) rather than linear functions. We study a version of the exponential weights algorithm and bound its regret in this setting. Under conditions on the eigendecay of the kernel we provide a sharp characterization of the regret for this algorithm. When we have polynomial eigendecay $\mu_j \le \mathcal{O}(j{-\beta})$, we find that the regret is bounded by $\mathcal{R}_n \le \mathcal{O}(n{\beta/(2(\beta-1))})$; while under the assumption of exponential eigendecay $\mu_j \le \mathcal{O}(e{-\beta j })$, we get an even tighter bound on the regret $\mathcal{R}_n \le \mathcal{O}(n{1/2}\log(n){1/2})$. We also study the full information setting when the underlying losses are kernel functions and present an adapted exponential weights algorithm and a conditional gradient descent algorithm.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Aldo Pacchiano (72 papers)
  2. Niladri S. Chatterji (21 papers)
  3. Peter L. Bartlett (86 papers)
Citations (18)

Summary

We haven't generated a summary for this paper yet.