MAGDiff: Covariate Data Set Shift Detection via Activation Graphs of Neural Networks

Published: 13 May 2024, Last Modified: 13 May 2024Accepted by TMLREveryoneRevisionsBibTeX
Abstract: Despite their successful application to a variety of tasks, neural networks remain limited, like other machine learning methods, by their sensitivity to shifts in the data: their performance can be severely impacted by differences in distribution between the data on which they were trained and that on which they are deployed. In this article, we propose a new family of representations, called MAGDiff, that we extract from any given neural network classifier and that allows for efficient covariate data shift detection without the need to train a new model dedicated to this task. These representations are computed by comparing the activation graphs of the neural network for samples belonging to the training distribution and to the target distribution, and yield powerful data- and task-adapted statistics for the two-sample tests commonly used for data set shift detection. We demonstrate this empirically by measuring the statistical powers of two-sample Kolmogorov-Smirnov (KS) tests on several different data sets and shift types, and showing that our novel representations induce significant improvements over a state-of-the-art baseline relying on the network output.
Submission Length: Regular submission (no more than 12 pages of main content)
Previous TMLR Submission Url:
Changes Since Last Submission: The font was wrong, due to an out-of-place \usepackage that escaped our attention despite our best efforts to comply with the template. We apologize for this oversight. The font should now be correct.
Supplementary Material: zip
Assigned Action Editor: ~Nadav_Cohen1
Submission Number: 2216