Bandits with Global Convex Constraints and ObjectiveOpen Website

2019 (modified: 08 Nov 2022)Oper. Res. 2019Readers: Everyone
Abstract: Multiarmed bandit (MAB) is a classic model for capturing the exploration–exploitation trade-off inherent in many sequential decision-making problems. The classic MAB framework, however, only allows...
0 Replies

Loading