Identifying Spurious Biases Early in Training through the Lens of Simplicity Bias

23 Sept 2023 (modified: 25 Mar 2024)ICLR 2024 Conference Withdrawn SubmissionEveryoneRevisionsBibTeX
Keywords: spurious correlations, simplicity bias
Abstract: Neural networks trained with (stochastic) gradient descent have an inductive bias towards learning simpler solutions. This makes them highly prone to learning simple \textit{spurious} features that are highly correlated with a label instead of the predictive but more complex core features. In this work, we show that, interestingly, the simplicity bias of gradient descent, can be leveraged to identify spurious correlations early in training. We provide theoretical insights on a two-layer neural network that subsets of data points where the spurious features strongly influence the label predictions are separable based on the model's output in the initial training iterations. We further show that if spurious features have a small enough noise-to-signal ratio, the network’s output on the majority of examples containing the spurious feature will be almost exclusively determined by the spurious features and will be nearly invariant to the core feature, leading to poor generalization performance for minority groups. Building on these findings, we propose SPARE, which separates groups with spurious features early in training and utilizes importance sampling to alleviate the spurious correlation by balancing the group sizes. Through rigorous experiments, we first establish SPARE's effectiveness in discovering spurious correlations in the Restricted ImageNet dataset. We then demonstrate that SPARE outperforms state-of-the-art methods by up to 5.6\% in worst-group accuracy, while being up to 12x faster.
Supplementary Material: zip
Primary Area: general machine learning (i.e., none of the above)
Code Of Ethics: I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics.
Submission Guidelines: I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2024/AuthorGuide.
Anonymous Url: I certify that there is no URL (e.g., github page) that could be used to find authors' identity.
No Acknowledgement Section: I certify that there is no acknowledgement section in this submission for double blind review.
Submission Number: 8137
Loading