Learning Curves for SGD on Structured FeaturesDownload PDF

29 Sept 2021, 00:35 (edited 13 Mar 2022)ICLR 2022 PosterReaders: Everyone
  • Keywords: Stochastic Gradient Descent, Generalization
  • Abstract: The generalization performance of a machine learning algorithm such as a neural network depends in a non-trivial way on the structure of the data distribution. To analyze the influence of data structure on test loss dynamics, we study an exactly solveable model of stochastic gradient descent (SGD) on the square loss which predicts test error when training on features with arbitrary covariance structure. We solve the theory exactly for both Gaussian features and arbitrary features and we show that the simpler Gaussian model accurately predicts test loss of nonlinear random-feature models and neural networks in the kernel regime trained with SGD on real datasets such as MNIST and CIFAR-10. We show that the optimal batch size at a fixed compute budget is typically small and depends on the feature correlation structure, demonstrating the computational benefits of SGD with small batch sizes. Lastly, we extend our theory to the more usual setting of stochastic gradient descent on a fixed subsampled training set, showing that both training and test error can be accurately predicted in our framework on real data.
  • One-sentence Summary: The average case test risk for stochastic gradient descent on mean square error is computed in terms of feature covariance structure.
  • Supplementary Material: zip
23 Replies