Overcoming the curse of dimensionality with Laplacian regularization in semi-supervised learningDownload PDF

21 May 2021, 20:41 (edited 12 Oct 2021)NeurIPS 2021 PosterReaders: Everyone
  • Keywords: semi-supervised learning, Laplacian, Laplacian regularization, weakly supervised learning, kernel methods, spectral filtering, partial labelling, diffusion, functional analysis, statistical learning, generalization bounds
  • TL;DR: Laplacian regularization the right way
  • Abstract: As annotations of data can be scarce in large-scale practical problems, leveraging unlabelled examples is one of the most important aspects of machine learning. This is the aim of semi-supervised learning. To benefit from the access to unlabelled data, it is natural to diffuse smoothly knowledge of labelled data to unlabelled one. This induces to the use of Laplacian regularization. Yet, current implementations of Laplacian regularization suffer from several drawbacks, notably the well-known curse of dimensionality. In this paper, we design a new class of algorithms overcoming this issue, unveiling a large body of spectral filtering methods. Additionally, we provide a statistical analysis showing that our estimators exhibit desirable behaviors. They are implemented through (reproducing) kernel methods, for which we provide realistic computational guidelines in order to make our method usable with large amounts of data.
  • Supplementary Material: pdf
  • Code Of Conduct: I certify that all co-authors of this work have read and commit to adhering to the NeurIPS Statement on Ethics, Fairness, Inclusivity, and Code of Conduct.
  • Code: https://github.com/VivienCabannes/partial_labelling
7 Replies