Papers
Topics
Authors
Recent
2000 character limit reached

An Asymptotically Optimal Policy for Uniform Bandits of Unknown Support

Published 8 May 2015 in stat.ML and cs.LG | (1505.01918v3)

Abstract: Consider the problem of a controller sampling sequentially from a finite number of $N \geq 2$ populations, specified by random variables $Xi_k$, $ i = 1,\ldots , N,$ and $k = 1, 2, \ldots$; where $Xi_k$ denotes the outcome from population $i$ the $k{th}$ time it is sampled. It is assumed that for each fixed $i$, ${ Xi_k }_{k \geq 1}$ is a sequence of i.i.d. uniform random variables over some interval $[a_i, b_i]$, with the support (i.e., $a_i, b_i$) unknown to the controller. The objective is to have a policy $\pi$ for deciding, based on available data, from which of the $N$ populations to sample from at any time $n=1,2,\ldots$ so as to maximize the expected sum of outcomes of $n$ samples or equivalently to minimize the regret due to lack on information of the parameters ${ a_i }$ and ${ b_i }$. In this paper, we present a simple inflated sample mean (ISM) type policy that is asymptotically optimal in the sense of its regret achieving the asymptotic lower bound of Burnetas and Katehakis (1996). Additionally, finite horizon regret bounds are given.

Citations (27)

Summary

We haven't generated a summary for this paper yet.

Whiteboard

Paper to Video (Beta)

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.