From Generalization to Inner Activations
Abstract: One of the side effects of deep learning models becoming increasingly large is the amount of data that they intermediately generate which is then often ignored. Researchers have used this information to create explainability methods which show importance relative to a specific input it relates to, but often the intermediate values are disregarded and mostly underutilized. Although this intermediate data can become increasingly unwieldily in size as the networks grow, the ability to monitor specific layers is a valuable tool that provides insight into how the model is learning as well as vague generalities about how the model performs overall (e.g. where are specific features begun to be extracted or where is noise filtered out).
Submission Full: zip
Blogpost Url: yml
ICLR Paper: https://arxiv.org/abs/1611.03530, https://arxiv.org/abs/1803.06959, http://arxiv.org/abs/1706.05806, https://arxiv.org/abs/1806.05759
2 Replies
Loading