How robust are pre-trained models to distribution shift? Download PDF

Published: 21 Jul 2022, Last Modified: 25 Nov 2024SCIS 2022 PosterReaders: Everyone
Keywords: spurious correlation, pre-trained models, distribution shift, unsupervised learning, self-supervised learning, auto-encoder, variational auto-encoder
TL;DR: We compare the robustness of supervised learning, self-supervised learning and auto-encoding models to distribution shift.
Abstract: The vulnerability of machine learning models to spurious correlations has mostly been discussed in the context of supervised learning (SL). However, there is a lack of insight on how spurious correlations affect the performance of popular self-supervised learning (SSL) and auto-encoder based models (AE). In this work, we shed light on this by evaluating the performance of these models on both real world and synthetic distribution shift datasets. Following observations that the linear head itself can be susceptible to spurious correlations, we develop a new evaluation scheme with the linear head trained on out-of-distribution (OOD) data, to isolate the performance of the pre-trained models from a potential bias of the linear head used for evaluation. With this new methodology, we show that SSL models are consistently more robust to distribution shifts and thus better at OOD generalisation than AE and SL models.
Confirmation: Yes
Community Implementations: [![CatalyzeX](/images/catalyzex_icon.svg) 1 code implementation](https://www.catalyzex.com/paper/how-robust-are-pre-trained-models-to/code)
0 Replies

Loading