Position: Rethinking Explainable Machine Learning as Applied Statistics

Published: 01 May 2025, Last Modified: 23 Jul 2025ICML 2025 Position Paper Track posterEveryoneRevisionsBibTeXCC BY-SA 4.0
TL;DR: We argue that explainable machine learning is applied statistics for high-dimensional functions.
Abstract: In the rapidly growing literature on explanation algorithms, it often remains unclear what precisely these algorithms are for and how they should be used. In this position paper, we argue for a novel and pragmatic perspective: Explainable machine learning needs to recognize its parallels with applied statistics. Concretely, explanations are statistics of high-dimensional functions, and we should think about them analogously to traditional statistical quantities. Among others, this implies that we must think carefully about the matter of interpretation, or how the explanations relate to intuitive questions that humans have about the world. The fact that this is scarcely being discussed in research papers is one of the main drawbacks of the current literature. Moving forward, the analogy between explainable machine learning and applied statistics provides a fruitful way for how research practices can be improved.
Lay Summary: Machine learning models increasingly influence important decisions, from loan approvals to medical diagnoses. This has led to growing interest in explainable machine learning — methods that aim to make model behavior transparent. However, after years of research, it still remains unclear what exactly it means to "explain" a model. We argue that the solution is surprisingly simple: explainable machine learning is statistics by another name. Just as statistics provides tools to analyze large datasets, explainable machine learning provides tools to analyze large models. While this sounds simple in hindsight, it's actually a fundamental shift in perspective. By treating model explanations as statistics, we can apply basic lessons from statistical practice: always specify what your tool measures, acknowledge its limitations, and ensure users understand how to interpret it. It is important to get this right because explainable machine learning is not only used in research but also in applications and policy contexts.
Verify Author Names: My co-authors have confirmed that their names are spelled correctly both on OpenReview and in the camera-ready PDF. (If needed, please update ‘Preferred Name’ in OpenReview to match the PDF.)
No Additional Revisions: I understand that after the May 29 deadline, the camera-ready submission cannot be revised before the conference. I have verified with all authors that they approve of this version.
Pdf Appendices: My camera-ready PDF file contains both the main text (not exceeding the page limits) and all appendices that I wish to include. I understand that any other supplementary material (e.g., separate files previously uploaded to OpenReview) will not be visible in the PMLR proceedings.
Latest Style File: I have compiled the camera ready paper with the latest ICML2025 style files <https://media.icml.cc/Conferences/ICML2025/Styles/icml2025.zip> and the compiled PDF includes an unnumbered Impact Statement section.
Paper Verification Code: ODU5M
Permissions Form: pdf
Primary Area: Model Understanding, Explainability, Interpretability, and Trust
Keywords: Explainability, Interpretability
Submission Number: 171
Loading