Consistent Sufficient Explanations and Minimal Local Rules for explaining the decision of any classifier or regressorDownload PDF

Published: 31 Oct 2022, 18:00, Last Modified: 11 Jan 2023, 11:26NeurIPS 2022 AcceptReaders: Everyone
Keywords: Interpretability, Trustworthy ML, Robust and Reliable ML, rule-based models, learning theory, random forests, explainable ai, consistency, tree-based models
Abstract: To explain the decision of any regression and classification model, we extend the notion of probabilistic sufficient explanations (P-SE). For each instance, this approach selects the minimal subset of features that is sufficient to yield the same prediction with high probability, while removing other features. The crux of P-SE is to compute the conditional probability of maintaining the same prediction. Therefore, we introduce an accurate and fast estimator of this probability via random Forests for any data $(\boldsymbol{X}, Y)$ and show its efficiency through a theoretical analysis of its consistency. As a consequence, we extend the P-SE to regression problems. In addition, we deal with non-discrete features, without learning the distribution of $\boldsymbol{X}$ nor having the model for making predictions. Finally, we introduce local rule-based explanations for regression/classification based on the P-SE and compare our approaches w.r.t other explainable AI methods. These methods are available as a Python Package.
Supplementary Material: pdf
16 Replies

Loading