2019 (modified: 08 Nov 2022)Oper. Res. 2019Readers: Everyone
Abstract:Multiarmed bandit (MAB) is a classic model for capturing the exploration–exploitation trade-off inherent in many sequential decision-making problems. The classic MAB framework, however, only allows...