Evaluating Groups of Features via Consistency, Contiguity, and Stability

Published: 19 Mar 2024, Last Modified: 01 Jun 2024Tiny Papers @ ICLR 2024 NotableEveryoneRevisionsBibTeXCC BY 4.0
Keywords: Feature Attribution, Explainability, Group Evaluation, Group Attribution
TL;DR: We study consistency, contiguity, and stability for groups of features and find that modern grouping methods are no better than patches.
Abstract: Feature attributions explain model predictions by assigning importance scores to input features. In high-dimensional data such as images, these scores are often assigned to groups of features. There are various strategies for creating these groups, ranging from simple patches to deep-learning-based algorithms. Which group should be used for explanation? We formally define three key criteria for interpretable groups of features: consistency, contiguity, and stability. We find that patch-based groups outperform groups created via modern segmentation tools.
Submission Number: 147
Loading