An explore-then-commit algorithm for submodular maximization under full-bandit feedbackDownload PDFOpen Website

Published: 01 Jan 2022, Last Modified: 27 Apr 2023UAI 2022Readers: Everyone
Abstract: We investigate the problem of combinatorial multi-armed bandits with stochastic submodular (in expectation) rewards and full-bandit feedback, where no extra information other than the reward of sel...
0 Replies

Loading