2021 (modified: 03 Nov 2022)Math. Oper. Res. 2021Readers: Everyone
Abstract:We consider an N-player multiarmed bandit game in which each player chooses one out of M arms for T turns. Each player has different expected rewards for the arms, and the instantaneous rewards are...