2019 (modified: 08 Nov 2022)AISTATS 2019Readers: Everyone
Abstract:Stochastic Gradient Descent (SGD) is a central tool in machine learning. We prove that SGD converges to zero loss, even with a fixed (non-vanishing) learning rate — in the special case of homogeneo...