We design and evaluate a new algorithm called FlexCompressSGD for training deep neural networks over distributed datasets via multiple workers and a central server. In FlexCompressSGD, all gradients transmitted between workers and the server are compressed, and the workers are allowed to flexibly choose a compressing method different from that of the server. This flexibility significantly helps reduce the communication cost from each worker to the server. We mathematically prove that FlexCompressSGD converges with convergence rate /\sqrt {MT}$ where $M$ is the number of distributed workers and $T$ is the number of training iterations. We experimentally demonstrate that FlexCompressSGD obtains competitive top-1 testing accuracy on the ImageNet dataset while being able to reduce more than 70% communication cost from each worker to the server when compared with the state-of-the-art.
Mendeley helps you to discover research relevant for your work.
CITATION STYLE
Phuong, T. T., & Phong, L. T. (2020). Distributed SGD with Flexible Gradient Compression. IEEE Access, 8, 64707–64717. https://doi.org/10.1109/ACCESS.2020.2984633