## On the Convex Behavior of Deep Neural Networks in Relation to the Layers' Width

17 May 2019, 14:25 (modified: 08 Jul 2022, 18:53)ICML Deep Phenomena 2019Readers: Everyone
Abstract: The Hessian of neural networks can be decomposed into a sum of two matrices: (i) the positive semidefinite generalized Gauss-Newton matrix \$G\$, and (ii) the matrix \$H\$ containing negative eigenvalues. We observe that for wider networks, minimizing the loss with the gradient descent optimization maneuvers through surfaces of positive curvatures at the start and end of training, and close to zero curvatures in between. In other words, it seems that during crucial parts of the training process, the Hessian in wide networks is dominated by the component \$G\$. To explain this phenomenon, we show that when initialized using common methodologies, the gradients of over-parameterized networks are approximately orthogonal to \$H\$, such that the curvature of the loss surface is strictly positive in the direction of the gradient.
1 Reply

Loading