FedDR – Randomized Douglas-Rachford Splitting Algorithms for Nonconvex Federated Composite OptimizationDownload PDF

21 May 2021, 20:41 (modified: 26 Oct 2021, 19:04)NeurIPS 2021 PosterReaders: Everyone
Keywords: Randomized Douglas-Rachford splitting method, nonconvex federated optimization, asynchronous algorithm
TL;DR: This paper proposes two novel randomized Douglas-Rachford splitting algorithms for nonconvex federated optimization
Abstract: We develop two new algorithms, called, FedDR and asyncFedDR, for solving a fundamental nonconvex composite optimization problem in federated learning. Our algorithms rely on a novel combination between a nonconvex Douglas-Rachford splitting method, randomized block-coordinate strategies, and asynchronous im- plementation. They can also handle convex regularizers. Unlike recent methods in the literature, e.g., FedSplit and FedPD, our algorithms update only a subset of users at each communication round, and possibly in an asynchronous manner, making them more practical. These new algorithms can handle statistical and sys- tem heterogeneity, which are the two main challenges in federated learning, while achieving the best known communication complexity. In fact, our new algorithms match the communication complexity lower bound up to a constant factor under standard assumptions. Our numerical experiments illustrate the advantages of our methods over existing algorithms on synthetic and real datasets.
Supplementary Material: pdf
Code Of Conduct: I certify that all co-authors of this work have read and commit to adhering to the NeurIPS Statement on Ethics, Fairness, Inclusivity, and Code of Conduct.
Code: https://github.com/unc-optimization/FedDR
16 Replies