Keywords: label noise; multi-class classification; learning theory; domain adaptation; minimax rate; self-supervised learning
Abstract: We establish a new theoretical framework for learning under multi-class, instance-dependent label noise. At the heart of our framework is the concept of \emph{relative signal strength} (RSS), which is a point-wise measure of noisiness. We use relative signal strength to establish matching upper and lower bounds for excess risk. Our theoretical findings reveal a surprising result: the extremely simple \emph{Noise Ignorant Empirical Risk Minimization (NI-ERM)} principle, which conducts empirical risk minimization as if no label noise exists, is minimax optimal. Finally, we translate these theoretical insights into practice: by using NI-ERM to fit a linear classifier on top of a frozen foundation model, we achieve state-of-the-art performance on the CIFAR-N data challenge.
Is Neurips Submission: Yes
Submission Number: 32
Loading