Gummadi, Ramki; Johari, Ramesh; Schmit, Sven; Yu, Jia Yuan - 2016
Much of the classical work on algorithms for multi-armed bandits focuses on rewards that are stationary over time. By contrast, we study multi-armed bandit (MAB) games, where the rewards obtained by an agent also depend on how many other agents choose the same arm (as might be the case in many...