Distributed Online Convex Optimization with Compressed CommunicationDownload PDF

Published: 31 Oct 2022, 18:00, Last Modified: 13 Jan 2023, 11:52NeurIPS 2022 AcceptReaders: Everyone
Keywords: distributed online optimization, compressed communication, no-regret algorithm
TL;DR: We design provably no-regret distributed online algorithms that work with compressors.
Abstract: We consider a distributed online convex optimization problem when streaming data are distributed among computing agents over a connected communication network. Since the data are high-dimensional or the network is large-scale, communication load can be a bottleneck for the efficiency of distributed algorithms. To tackle this bottleneck, we apply the state-of-art data compression scheme to the fundamental GD-based distributed online algorithms. Three algorithms with difference-compressed communication are proposed for full information feedback (DC-DOGD), one-point bandit feedback (DC-DOBD), and two-point bandit feedback (DC-DO2BD), respectively. We obtain regret bounds explicitly in terms of time horizon, compression ratio, decision dimension, agent number, and network parameters. Our algorithms are proved to be no-regret and match the same regret bounds, w.r.t. time horizon, with their uncompressed versions for both convex and strongly convex losses. Numerical experiments are given to validate the theoretical findings and illustrate that the proposed algorithms can effectively reduce the total transmitted bits for distributed online training compared with the uncompressed baseline.
Supplementary Material: pdf
17 Replies