Formulating Generalizable and Non-Generalizable Interactions in DNNs

08 Sept 2025 (modified: 11 Feb 2026)Submitted to ICLR 2026EveryoneRevisionsBibTeXCC BY 4.0
Keywords: Overfitting; Generalization; Deep Learning Theory
TL;DR: This paper disentangles the distributions of generalizable interactions and non-generalizable interactions from a trained DNN.
Abstract: This paper aims to analyze the generalization power of deep neural networks (DNNs) from the perspective of interactions. Unlike previous analyses of a DNN’s generalization power in a high-dimensional feature space, we find that the generalization power of a DNN can be explained as the generalization power of the interactions. We find that generalizable interactions follow a decay-shaped distribution, while non-generalizable interactions follow a spindle-shaped distribution. Furthermore, we develop a method to disentangle these two types of interactions in a DNN. We have verified that the theoretically disentangled distributions of generalizable interactions and non-generalizable interactions can well match the real distributions in experiments.
Supplementary Material: zip
Primary Area: interpretability and explainable AI
Submission Number: 3161
Loading