Beyond Tikhonov: faster learning with self-concordant losses, via iterative regularizationDownload PDF

May 21, 2021 (edited Oct 25, 2021)NeurIPS 2021 SpotlightReaders: Everyone
  • Keywords: Kernel methods, learning theory, self concordance, iterated tikhonov, proximal point
  • TL;DR: The iterative Thikonov regularization scheme achieves optimal sample complexity on self concordant losses.
  • Abstract: The theory of spectral filtering is a remarkable tool to understand the statistical properties of learning with kernels. For least squares, it allows to derive various regularization schemes that yield faster convergence rates of the excess risk than with Tikhonov regularization. This is typically achieved by leveraging classical assumptions called source and capacity conditions, which characterize the difficulty of the learning task. In order to understand estimators derived from other loss functions, Marteau-Ferey et al. have extended the theory of Tikhonov regularization to generalized self concordant loss functions (GSC), which contain, e.g., the logistic loss. In this paper, we go a step further and show that fast and optimal rates can be achieved for GSC by using the iterated Tikhonov regularization scheme, which is intrinsically related to the proximal point method in optimization, and overcomes the limitation of the classical Tikhonov regularization.
  • Supplementary Material: pdf
  • Code Of Conduct: I certify that all co-authors of this work have read and commit to adhering to the NeurIPS Statement on Ethics, Fairness, Inclusivity, and Code of Conduct.
  • Code: zip
10 Replies

Loading