Trading Redundancy for Communication: Speeding up Distributed SGD for Non-convex OptimizationDownload PDFOpen Website

2019 (modified: 11 Nov 2022)ICML 2019Readers: Everyone
Abstract: Communication overhead is one of the key challenges that hinders the scalability of distributed optimization algorithms to train large neural networks. In recent years, there has been a great deal ...
0 Replies

Loading