From Generalization to Inner Activations

Anonymous

17 Jan 2022 (modified: 05 May 2023)Submitted to BT@ICLR2022Readers: Everyone
Abstract: One of the side effects of deep learning models becoming increasingly large is the amount of data that they intermediately generate which is then often ignored. Researchers have used this information to create explainability methods which show importance relative to a specific input it relates to, but often the intermediate values are disregarded and mostly underutilized. Although this intermediate data can become increasingly unwieldily in size as the networks grow, the ability to monitor specific layers is a valuable tool that provides insight into how the model is learning as well as vague generalities about how the model performs overall (e.g. where are specific features begun to be extracted or where is noise filtered out).
Submission Full: zip
Blogpost Url: yml
ICLR Paper: https://arxiv.org/abs/1611.03530, https://arxiv.org/abs/1803.06959, http://arxiv.org/abs/1706.05806, https://arxiv.org/abs/1806.05759
2 Replies

Loading