Keywords: Generalization, Input corruption, Noise Injection, Deep Learning
TL;DR: We explore "Memorization vs. Generalization" in DNNs by injecting noise into existing networks via optimizable weights. We relate the non-trivial phases experienced by the system during training to the loss curvature.
Abstract: We propose a new method to probe the learning mechanism of Deep Neural Networks (DNN) by perturbing the system using Noise Injection Nodes (NINs). These nodes inject uncorrelated noise via additional optimizable weights to existing feed-forward network architectures, without changing the optimization algorithm. We find that the system displays distinct phases during training, dictated by the scale of injected noise. We first derive expressions for the dynamics of the network and utilize a simple linear model as a test case. We find that in some cases, the evolution of the noise nodes is similar to that of the unperturbed loss, thus indicating the possibility of using NINs to learn more about the full system in the future.
0 Replies
Loading