Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Bandits Under The Influence (Extended Version) (2009.10135v1)

Published 21 Sep 2020 in cs.LG, cs.DB, and stat.ML

Abstract: Recommender systems should adapt to user interests as the latter evolve. A prevalent cause for the evolution of user interests is the influence of their social circle. In general, when the interests are not known, online algorithms that explore the recommendation space while also exploiting observed preferences are preferable. We present online recommendation algorithms rooted in the linear multi-armed bandit literature. Our bandit algorithms are tailored precisely to recommendation scenarios where user interests evolve under social influence. In particular, we show that our adaptations of the classic LinREL and Thompson Sampling algorithms maintain the same asymptotic regret bounds as in the non-social case. We validate our approach experimentally using both synthetic and real datasets.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Silviu Maniu (11 papers)
  2. Stratis Ioannidis (67 papers)
  3. Bogdan Cautis (14 papers)
Citations (2)

Summary

We haven't generated a summary for this paper yet.