Hiding in a Plain Sight: Out-of-Distribution Data in the Logit Space Embeddings

Published: 10 Oct 2024, Last Modified: 09 Nov 2024SciForDL PosterEveryoneRevisionsBibTeXCC BY 4.0
TL;DR: This study reveals that out-of-distribution (OOD) data concentrate near the center of the logit space, while in-distribution (ID) data forms distant class-wise clusters.
Abstract: Out-of-distribution (OOD) data are detrimental to the performance of deep learning (DL) classifiers, leading to extensive research focused on their detection. Current state-of-the-art OOD detection methods employ a scoring technique designed to assign lower scores to OOD samples compared to in-distribution (ID) ones. Nevertheless, these approaches lack foresight into the configuration of OOD and ID data within the latent space, instead making an implicit assumption regarding their inherent separation. As a result, most OOD detection methods result in complicated and hard-to-validate scoring techniques. This study conducts a thorough analysis of the logit embedding landscape, revealing that both ID and OOD data exhibit a distinct trend. Specifically, we demonstrate that OOD data tends to reside near to the center of the logit space. In contrast, ID data tends to be situated farther from the center, predominantly in the positive regions of the logit space, thus forming class-wise clusters along the orthogonal axes that span the logit space. This study highlights the critical role of the DL classifier in differentiating between ID and OOD logits.
Style Files: I have used the style files.
Debunking Challenge: This submission is an entry to the debunking challenge.
Submission Number: 37
Loading

OpenReview is a long-term project to advance science through improved peer review with legal nonprofit status. We gratefully acknowledge the support of the OpenReview Sponsors. © 2025 OpenReview