Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Adversarial Attacks on Linear Contextual Bandits (2002.03839v3)

Published 10 Feb 2020 in cs.LG and stat.ML

Abstract: Contextual bandit algorithms are applied in a wide range of domains, from advertising to recommender systems, from clinical trials to education. In many of these domains, malicious agents may have incentives to attack the bandit algorithm to induce it to perform a desired behavior. For instance, an unscrupulous ad publisher may try to increase their own revenue at the expense of the advertisers; a seller may want to increase the exposure of their products, or thwart a competitor's advertising campaign. In this paper, we study several attack scenarios and show that a malicious agent can force a linear contextual bandit algorithm to pull any desired arm $T - o(T)$ times over a horizon of $T$ steps, while applying adversarial modifications to either rewards or contexts that only grow logarithmically as $O(\log T)$. We also investigate the case when a malicious agent is interested in affecting the behavior of the bandit algorithm in a single context (e.g., a specific user). We first provide sufficient conditions for the feasibility of the attack and we then propose an efficient algorithm to perform the attack. We validate our theoretical results on experiments performed on both synthetic and real-world datasets.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Evrard Garcelon (13 papers)
  2. Laurent Meunier (18 papers)
  3. Jean Tarbouriech (10 papers)
  4. Olivier Teytaud (45 papers)
  5. Alessandro Lazaric (78 papers)
  6. Matteo Pirotta (45 papers)
  7. Baptiste Roziere (13 papers)
Citations (48)

Summary

We haven't generated a summary for this paper yet.