Explanations of Black-Box Models based on Directional Feature InteractionsDownload PDF

29 Sept 2021, 00:32 (edited 14 Mar 2022)ICLR 2022 SpotlightReaders: Everyone
  • Keywords: Explainability, Shapley values, Interpretability, Directional interaction, feature interaction
  • Abstract: As machine learning algorithms are deployed ubiquitously to a variety of domains, it is imperative to make these often black-box models transparent. Several recent works explain black-box models by capturing the most influential features for prediction per instance; such explanation methods are univariate, as they characterize importance per feature. We extend univariate explanation to a higher-order; this enhances explainability, as bivariate methods can capture feature interactions in black-box models, represented as a directed graph. Analyzing this graph enables us to discover groups of features that are equally important (i.e., interchangeable), while the notion of directionality allows us to identify the most influential features. We apply our bivariate method on Shapley value explanations, and experimentally demonstrate the ability of directional explanations to discover feature interactions. We show the superiority of our method against state-of-the-art on CIFAR10, IMDB, Census, Divorce, Drug, and gene data.
  • One-sentence Summary: We introduce a bivariate explainer to explain directional feature interactions in black box models.
  • Supplementary Material: zip
16 Replies

Loading