Identifying and interpreting tuning dimensions in deep networksDownload PDF

Published: 03 Nov 2020, Last Modified: 22 Oct 2023SVRHM@NeurIPS PosterReaders: Everyone
Keywords: Explainable artificial intelligence, interpretability, deep learning, computer vision, neuroscience
TL;DR: This work contributes an unsupervised framework for identifying and interpreting tuning dimensions in deep networks.
Abstract: In neuroscience, a tuning dimension is a stimulus attribute that accounts for much of the activation variance of a group of neurons. These are commonly used to decipher the responses of such groups. While researchers have attempted to manually identify an analogue to these tuning dimensions in deep neural networks, we are unaware of an automatic way to discover them. This work contributes an unsupervised framework for identifying and interpreting "tuning dimensions" in deep networks. Our method correctly identifies the tuning dimensions of a synthetic Gabor filter bank and tuning dimensions of the first two layers of InceptionV1 trained on ImageNet.
Community Implementations: [![CatalyzeX](/images/catalyzex_icon.svg) 1 code implementation](https://www.catalyzex.com/paper/arxiv:2011.03043/code)
5 Replies

Loading