Surrogate Regret Bounds for Polyhedral LossesDownload PDF

21 May 2021, 20:52 (modified: 15 Jan 2022, 20:05)NeurIPS 2021 PosterReaders: Everyone
Keywords: surrogate regret bounds, excess risk bounds, polyhedral losses, calibration, property elicitation
TL;DR: Any polyhedral surrogate loss achieves a linear surrogate regret bound, while "non-polyhedral" losses do not.
Abstract: Surrogate risk minimization is an ubiquitous paradigm in supervised machine learning, wherein a target problem is solved by minimizing a surrogate loss on a dataset. Surrogate regret bounds, also called excess risk bounds, are a common tool to prove generalization rates for surrogate risk minimization. While surrogate regret bounds have been developed for certain classes of loss functions, such as proper losses, general results are relatively sparse. We provide two general results. The first gives a linear surrogate regret bound for any polyhedral (piecewise-linear and convex) surrogate, meaning that surrogate generalization rates translate directly to target rates. The second shows that for sufficiently non-polyhedral surrogates, the regret bound is a square root, meaning fast surrogate generalization rates translate to slow rates for the target. Together, these results suggest polyhedral surrogates are optimal in many cases.
Supplementary Material: pdf
Code Of Conduct: I certify that all co-authors of this work have read and commit to adhering to the NeurIPS Statement on Ethics, Fairness, Inclusivity, and Code of Conduct.
8 Replies

Loading