Optimal Locally Private Nonparametric Classification with Public Data

Published: 30 May 2024, Last Modified: 20 Jul 2024Journal of Machine Learning ResearchEveryoneCC BY 4.0
Abstract: In this work, we investigate the problem of public data assisted non-interactive Lo- cal Differentially Private (LDP) learning with a focus on non-parametric classification. Under the posterior drift assumption, we for the first time derive the mini-max optimal convergence rate with LDP constraint. Then, we present a novel approach, the locally differentially private classification tree, which attains the mini-max optimal convergence rate. Furthermore, we design a data-driven pruning procedure that avoids parameter tun- ing and provides a fast converging estimator. Comprehensive experiments conducted on synthetic and real data sets show the superior performance of our proposed methods. Both our theoretical and experimental findings demonstrate the effectiveness of public data com- pared to private data, which leads to practical suggestions for prioritizing non-private data collection.
Loading