Towards better understanding of gradient-based attribution methods for Deep Neural NetworksDownload PDF

15 Feb 2018 (modified: 14 Oct 2024)ICLR 2018 Conference Blind SubmissionReaders: Everyone
Abstract: Understanding the flow of information in Deep Neural Networks (DNNs) is a challenging problem that has gain increasing attention over the last few years. While several methods have been proposed to explain network predictions, there have been only a few attempts to compare them from a theoretical perspective. What is more, no exhaustive empirical comparison has been performed in the past. In this work we analyze four gradient-based attribution methods and formally prove conditions of equivalence and approximation between them. By reformulating two of these methods, we construct a unified framework which enables a direct comparison, as well as an easier implementation. Finally, we propose a novel evaluation metric, called Sensitivity-n and test the gradient-based attribution methods alongside with a simple perturbation-based attribution method on several datasets in the domains of image and text classification, using various network architectures.
TL;DR: Four existing backpropagation-based attribution methods are fundamentally similar. How to assess it?
Keywords: Deep Neural Networks, Attribution methods, Theory of deep learning
Code: [![github](/images/github_icon.svg) kundajelab/deeplift](https://github.com/kundajelab/deeplift) + [![Papers with Code](/images/pwc_icon.svg) 1 community implementation](https://paperswithcode.com/paper/?openreview=Sy21R9JAW)
Data: [CIFAR-10](https://paperswithcode.com/dataset/cifar-10), [IMDb Movie Reviews](https://paperswithcode.com/dataset/imdb-movie-reviews)
Community Implementations: [![CatalyzeX](/images/catalyzex_icon.svg) 1 code implementation](https://www.catalyzex.com/paper/towards-better-understanding-of-gradient/code)
8 Replies

Loading