Statistical Significance of Feature Importance Rankings

Published: 07 May 2025, Last Modified: 13 Jun 2025UAI 2025 PosterEveryoneRevisionsBibTeXCC BY 4.0
Keywords: Interpretability, Stability, Hypothesis Testing, Statistics, Ranking, Feature Importance, Attributions, Explainability
TL;DR: We guarantee the stability of feature importance rankings, both in retrospect and with novel top-K algorithms.
Abstract: Feature importance scores are ubiquitous tools for understanding the predictions of machine learning models. However, many popular attribution methods suffer from high instability due to random sampling. Leveraging novel ideas from hypothesis testing, we devise techniques that ensure the most important features are correct with high-probability guarantees. These are capable of assessing both the set of $K$ top-ranked features as well as the order of its elements. Given local or global importance scores, we demonstrate how to retrospectively verify the stability of the highest ranks. We then introduce two efficient sampling algorithms that identify the $K$ most important features, perhaps in order, with probability at least $1-\alpha$. The theoretical justification for these procedures is validated empirically on SHAP and LIME.
Supplementary Material: zip
Latex Source Code: zip
Code Link: https://github.com/jeremy-goldwasser/feature-rankings
Signed PMLR Licence Agreement: pdf
Readers: auai.org/UAI/2025/Conference, auai.org/UAI/2025/Conference/Area_Chairs, auai.org/UAI/2025/Conference/Reviewers, auai.org/UAI/2025/Conference/Submission76/Authors, auai.org/UAI/2025/Conference/Submission76/Reproducibility_Reviewers
Submission Number: 76
Loading