Adversarial Regression with Doubly Non-negative Weighting MatricesDownload PDF

May 21, 2021 (edited Oct 25, 2021)NeurIPS 2021 PosterReaders: Everyone
  • Keywords: adversarial regression, kernel weighted regression, covariate perturbation, distributional robust, doubly non-negative matrices, Bures-Wasserstein, Log-determinant divergence, low-sample-size problem
  • TL;DR: Adversarial Regression with Doubly Non-negative Weighting Matrices
  • Abstract: Many machine learning tasks that involve predicting an output response can be solved by training a weighted regression model. Unfortunately, the predictive power of this type of models may severely deteriorate under low sample sizes or under covariate perturbations. Reweighting the training samples has aroused as an effective mitigation strategy to these problems. In this paper, we propose a novel and coherent scheme for kernel-reweighted regression by reparametrizing the sample weights using a doubly non-negative matrix. When the weighting matrix is confined in an uncertainty set using either the log-determinant divergence or the Bures-Wasserstein distance, we show that the adversarially reweighted estimate can be solved efficiently using first-order methods. Numerical experiments show that our reweighting strategy delivers promising results on numerous datasets.
  • Supplementary Material: pdf
  • Code Of Conduct: I certify that all co-authors of this work have read and commit to adhering to the NeurIPS Statement on Ethics, Fairness, Inclusivity, and Code of Conduct.
  • Code: https://github.com/lttam/Adversarial-Regression
13 Replies

Loading