Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Empirical Bayes Regret Minimization (1904.02664v4)

Published 4 Apr 2019 in cs.LG and stat.ML

Abstract: Most bandit algorithm designs are purely theoretical. Therefore, they have strong regret guarantees, but also are often too conservative in practice. In this work, we pioneer the idea of algorithm design by minimizing the empirical Bayes regret, the average regret over problem instances sampled from a known distribution. We focus on a tractable instance of this problem, the confidence interval and posterior width tuning, and propose an efficient algorithm for solving it. The tuning algorithm is analyzed and evaluated in multi-armed, linear, and generalized linear bandits. We report several-fold reductions in Bayes regret for state-of-the-art bandit algorithms, simply by optimizing over a small sample from a distribution.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Branislav Kveton (98 papers)
  2. Ofer Meshi (14 papers)
  3. Martin Mladenov (22 papers)
  4. Csaba Szepesvari (157 papers)
  5. Chih-wei Hsu (14 papers)
Citations (12)

Summary

We haven't generated a summary for this paper yet.