Keywords: linear bandits, non-i.i.d, online learning
Abstract: We study the linear stochastic bandit problem, relaxing the standard i.i.d. assumption on the observation noise.
As an alternative to this restrictive assumption, we allow the noise terms across rounds to be sub-Gaussian but
interdependent, with dependencies that decay over time. To address this setting, we develop new confidence sequences
using a recently introduced reduction scheme to sequential probability assignment, and use these to derive a bandit
algorithm based on the principle of optimism in the face of uncertainty. We provide regret bounds for the
resulting algorithm, expressed in terms of the decay rate of the strength of dependence between observations. Among
other results, we show that our bounds recover the standard rates up to a factor of the mixing time for geometrically
mixing observation noise.
Supplementary Material: zip
Primary Area: Theory (e.g., control theory, learning theory, algorithmic game theory)
Submission Number: 12202
Loading