Convergence of Stochastic Gradient Langevin Dynamics in the Lazy Training Regime

Published: 01 Mar 2026, Last Modified: 01 Mar 2026Accepted by TMLREveryoneRevisionsBibTeXCC BY 4.0
Abstract: Continuous-time models provide important insights into the training dynamics of optimization algorithms in deep learning. In this work, we establish a non-asymptotic convergence analysis of stochastic gradient Langevin dynamics (SGLD), which is an Itô stochastic differential equation (SDE) approximation of stochastic gradient descent in continuous time, in the lazy training regime. We show that, under regularity conditions on the Hessian of the loss function, SGLD with multiplicative and state-dependent noise (i) yields a non-degenerate kernel throughout the training process with high probability, and (ii) achieves exponential convergence to the empirical risk minimizer in expectation, and we establish finite-time and finite-width bounds on the optimality gap. We corroborate our theoretical findings with numerical examples in the regression setting.
Submission Type: Regular submission (no more than 12 pages of main content)
Changes Since Last Submission: We identified some typographical issues and corrected them.
Assigned Action Editor: ~Alain_Durmus1
Submission Number: 6291
Loading