A Practical Algorithm for Multiplayer Bandits when Arm Means Vary Among Players

Research output: Contribution to journalConference articlepeer-review

Abstract

We study a multiplayer stochastic multi-armed bandit problem in which players cannot communicate, and if two or more players pull the same arm, a collision occurs and the involved players receive zero reward. We consider the challenging heterogeneous setting, in which different arms may have different means for different players, and propose a new and efficient algorithm that combines the idea of leveraging forced collisions for implicit communication and that of performing matching eliminations. We present a finite-time analysis of our algorithm, giving the first sublinear minimax regret bound for this problem, and prove that if the optimal assignment of players to arms is unique, our algorithm attains the optimal O(ln(T)) regret, solving an open question raised at NeurIPS 2018 by Bistritz and Leshem (2018).

Original languageEnglish
Pages (from-to)1211-1221
Number of pages11
JournalProceedings of Machine Learning Research
Volume108
Publication statusPublished - 1 Jan 2020
Event23rd International Conference on Artificial Intelligence and Statistics, AISTATS 2020 - Virtual, Online
Duration: 26 Aug 202028 Aug 2020

Fingerprint

Dive into the research topics of 'A Practical Algorithm for Multiplayer Bandits when Arm Means Vary Among Players'. Together they form a unique fingerprint.

Cite this