Enhancing Graph Learning Interpretability through Modulating Cluster Information Flow

Published: 26 Jan 2026, Last Modified: 27 Jan 2026Pattern RecognitionEveryoneCC BY 4.0
Abstract: Interpretable graph learning is essential for scientific applications that depend on learning models to extract reliable insights from graph-structured data. Recent efforts to explain GNN predictions focus on identifying vital substructures, such as subgraphs. However, existing approaches tend to misclassify the neighboring irrelevant nodes as part of the vital subgraphs. To address this, we propose Cluster Information Flow Graph Neural Networks (CIFlow-GNN), a built-in model-level method that provides accurate interpretable subgraph explanations by modulating the cluster information flow.  CIFlow-GNN incorporates two modules, i.e., the graph clustering module and the cluster prototype module. The graph clustering module partitions the nodes according to their connectivity in the graph topology and their similarity in cluster features. Specifically, we introduce a cluster feature loss to regulate information flow at the cluster level. We prove that the proposed cluster feature loss is a lower bound of the InfoNCE loss. Optimizing the cluster feature loss reduces the mutual information among clusters and achieves the modulation of cluster information flow. Subsequently, the graph prototype module uses prototypes as a bridge to select important clusters as vital subgraphs by integrating information across all graphs. To ensure accurate correspondence between clusters and prototypes, we further modulate the cluster information flow at the prototype level. Experimental studies on both synthetic and real world datasets demonstrate that our proposed CIFlow-GNN can identify vital subgraphs effectively and efficiently.
Loading