Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Productization Challenges of Contextual Multi-Armed Bandits (1907.04884v1)

Published 10 Jul 2019 in cs.IR and cs.LG

Abstract: Contextual Multi-Armed Bandits is a well-known and accepted online optimization algorithm, that is used in many Web experiences to tailor content or presentation to users' traffic. Much has been published on theoretical guarantees (e.g. regret bounds) of proposed algorithmic variants, but relatively little attention has been devoted to the challenges encountered while productizing contextual bandits schemes in large scale settings. This work enumerates several productization challenges we encountered while leveraging contextual bandits for two concrete use cases at scale. We discuss how to (1) determine the context (engineer the features) that model the bandit arms; (2) sanity check the health of the optimization process; (3) evaluate the process in an offline manner; (4) add potential actions (arms) on the fly to a running process; (5) subject the decision process to constraints; and (6) iteratively improve the online learning algorithm. For each such challenge, we explain the issue, provide our approach, and relate to prior art where applicable.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. David Abensur (1 paper)
  2. Ivan Balashov (2 papers)
  3. Shaked Bar (2 papers)
  4. Ronny Lempel (7 papers)
  5. Nurit Moscovici (1 paper)
  6. Ilan Orlov (2 papers)
  7. Danny Rosenstein (2 papers)
  8. Ido Tamir (1 paper)
Citations (2)

Summary

We haven't generated a summary for this paper yet.