Keywords: Grokking, feature learning, training dynamics, Neural Tangent Kernel
Abstract: In this paper, we analyze the phenomenon of Grokking in a sparse parity task trained with Deep
Neural Networks through the lens of feature learning. In particular, we analyze the evolution of the
Neural Tangent Kernel (NTK) matrix. We show that during the initial overfitting phase, the NTK’s
eigenfunctions are not aligned with the predictive input features. On the other hand, at a later stage
the NTK’s top eigenfunctions evolve to focus on the features of interest, which corresponds to the
onset of the delayed generalization typically observed in Grokking. Our experiments can be viewed
as a mechanistic interpretation of feature learning during training through the NTK eigenfunctions’
evolution.
Student Paper: Yes
Submission Number: 72
Loading