Stochastic Gradient Descent on Separable Data: Exact Convergence with a Fixed Learning RateDownload PDFOpen Website

2019 (modified: 08 Nov 2022)AISTATS 2019Readers: Everyone
Abstract: Stochastic Gradient Descent (SGD) is a central tool in machine learning. We prove that SGD converges to zero loss, even with a fixed (non-vanishing) learning rate — in the special case of homogeneo...
0 Replies

Loading