Out-of-Distribution Failure through the Lens of Labeling Mechanisms: An Information Theoretic ApproachDownload PDF

28 May 2022, 15:03 (modified: 21 Jul 2022, 01:30)SCIS 2022 PosterReaders: Everyone
Keywords: Out-of-distribution generalization, domain generalization, generalization bound
TL;DR: We deploy multiple labels for a datapoint to better disentangle and distinguish between the existing correlations in an environment, and offers a remedy which significantly boosts the performance of models over OOD test cases.
Abstract: Machine learning models typically fail in deployment environments where the distribution of data does not perfectly match that of the training domains. This phenomenon is believed to stem from networks' failure to capture the invariant features that generalize to unseen domains. However, we attribute this phenomenon to the limitations that the labeling mechanism employed by humans imposes on the learning algorithm. We conjecture that providing multiple labels for each datapoint where each could describe the existence of particular objects/concepts on the data point, decreases the risk of capturing non-generalizable correlations by the model. We theoretically show that learning over a multi-label regime, where $K$ labels for each data point are present, tightens the expected generalization gap by a factor of $1/\sqrt{K}$ compared to a similar case where only one label for each data point is in hand. Also, we show that learning under this regime is much more sample efficient and requires a fraction of training data to provide competitive results.
Confirmation: Yes
0 Replies