Better Rates for Private Linear Regression in the Proportional Regime via Aggressive Clipping

Published: 09 Jun 2025, Last Modified: 06 Jul 2025HiLD at ICML 2025 PosterEveryoneRevisionsBibTeXCC BY 4.0
Keywords: linear regression, differential privacy, proportional regime, gradient clipping, DP-SGD, deterministic equivalent
TL;DR: We provide better rates for DP linear regression, via a DP-SGD algorithm with a clipping constant of the same order of the gradients. We analyze the test risk via deterministic ODEs, that identify optimal hyper-parameters and learning rate schedules.
Abstract: Differentially private (DP) linear regression has received significant attention in the recent theoretical literature, with several works aimed at obtaining improved error rates. A common approach is to set the clipping constant much larger than the expected norm of the per-sample gradients. While simplifying the analysis, this is however in sharp contrast with what empirical evidence suggests to optimize performance. Our work bridges this gap between theory and practice: we provide sharper rates for DP stochastic gradient descent (DP-SGD) by crucially operating in a regime where clipping happens frequently. Specifically, we consider the setting where the data is multivariate Gaussian, the number of training samples $n$ is proportional to the input dimension $d$, and the algorithm guarantees constant-order zero concentrated DP. Our method relies on establishing a deterministic equivalent for the trajectory of DP-SGD in terms of a family of ordinary differential equations (ODEs). As a consequence, the risk of DP-SGD is bounded between two ODEs, with upper and lower bounds matching for isotropic data. By studying these ODEs when $n / d$ is large enough, we demonstrate the optimality of aggressive clipping, and we uncover the benefits of decaying learning rate and private noise scheduling.
Student Paper: Yes
Submission Number: 38
Loading