Generalization Bounds for (Wasserstein) Robust OptimizationDownload PDF

21 May 2021, 20:47 (edited 28 Jan 2022)NeurIPS 2021 PosterReaders: Everyone
  • Keywords: Distributionally robust optimization, Lipschitz regularization, gradient regularization, generalization bound
  • TL;DR: We derive finite-sample performance guarantees for robust optimization, Wasserstein roust optimization and their associated Lipschitz/gradient regularization under rather general conditions.
  • Abstract: (Distributionally) robust optimization has gained momentum in machine learning community recently, due to its promising applications in developing generalizable learning paradigms. In this paper, we derive generalization bounds for robust optimization and Wasserstein robust optimization for Lipschitz and piecewise Hölder smooth loss functions under both stochastic and adversarial setting, assuming that the underlying data distribution satisfies transportation-information inequalities. The proofs are built on new generalization bounds for variation regularization (such as Lipschitz or gradient regularization) and its connection with robustness.
  • Supplementary Material: pdf
  • Code Of Conduct: I certify that all co-authors of this work have read and commit to adhering to the NeurIPS Statement on Ethics, Fairness, Inclusivity, and Code of Conduct.
15 Replies