Risk Minimization from Adaptively Collected Data: Guarantees for Supervised and Policy LearningDownload PDF

21 May 2021, 20:51 (edited 26 Oct 2021)NeurIPS 2021 PosterReaders: Everyone
  • Keywords: empirical risk minimization, maximal inequality, metric entropy, fast rates, policy learning, empirical process theory, adaptively collected data
  • TL;DR: We provide first-of-their-kind generalization guarantees and fast convergence rates for importance sampling weighted empirical risk minimization from adaptively collected data, such as data collected by a contextual bandit algorithm.
  • Abstract: Empirical risk minimization (ERM) is the workhorse of machine learning, whether for classification and regression or for off-policy policy learning, but its model-agnostic guarantees can fail when we use adaptively collected data, such as the result of running a contextual bandit algorithm. We study a generic importance sampling weighted ERM algorithm for using adaptively collected data to minimize the average of a loss function over a hypothesis class and provide first-of-their-kind generalization guarantees and fast convergence rates. Our results are based on a new maximal inequality that carefully leverages the importance sampling structure to obtain rates with the good dependence on the exploration rate in the data. For regression, we provide fast rates that leverage the strong convexity of squared-error loss. For policy learning, we provide regret guarantees that close an open gap in the existing literature whenever exploration decays to zero, as is the case for bandit-collected data. An empirical investigation validates our theory.
  • Supplementary Material: pdf
  • Code Of Conduct: I certify that all co-authors of this work have read and commit to adhering to the NeurIPS Statement on Ethics, Fairness, Inclusivity, and Code of Conduct.
  • Code: https://github.com/mdimakopoulou/adaptive-erm
10 Replies