Abstract: There is a growing concern that the recent progress made in
AI, especially regarding the predictive competence of deep
learning models, will be undermined by a failure to properly
explain their operation and outputs. In response to this disquiet, counterfactual explanations have become very popular
in eXplainable AI (XAI) due to their asserted computational,
psychological, and legal benefits. In contrast however, semifactuals (which appear to be equally useful) have surprisingly received no attention. Most counterfactual methods address tabular rather than image data, partly because the nondiscrete nature of images makes good counterfactuals difficult to define; indeed, generating plausible counterfactual images which lie on the data manifold is also problematic. This
paper advances a novel method for generating plausible counterfactuals and semi-factuals for black-box CNN classifiers
doing computer vision. The present method, called PlausIble Exceptionality-based Contrastive Explanations (PIECE),
modifies all “exceptional” features in a test image to be “normal” from the perspective of the counterfactual class, to generate plausible counterfactual images. Two controlled experiments compare this method to others in the literature, showing that PIECE generates highly plausible counterfactuals
(and the best semi-factuals) on several benchmark measures.
0 Replies
Loading