Sparse-SignSGD with Majority Vote for Communication-Efficient Distributed Learning

02/15/2023
by   Chanho Park, et al.
0

The training efficiency of complex deep learning models can be significantly improved through the use of distributed optimization. However, this process is often hindered by a large amount of communication cost between workers and a parameter server during iterations. To address this bottleneck, in this paper, we present a new communication-efficient algorithm that offers the synergistic benefits of both sparsification and sign quantization, called S^3GD-MV. The workers in S^3GD-MV select the top-K magnitude components of their local gradient vector and only send the signs of these components to the server. The server then aggregates the signs and returns the results via a majority vote rule. Our analysis shows that, under certain mild conditions, S^3GD-MV can converge at the same rate as signSGD while significantly reducing communication costs, if the sparsification parameter K is properly chosen based on the number of workers and the size of the deep learning model. Experimental results using both independent and identically distributed (IID) and non-IID datasets demonstrate that the S^3GD-MV attains higher accuracy than signSGD, significantly reducing communication costs. These findings highlight the potential of S^3GD-MV as a promising solution for communication-efficient distributed optimization in deep learning.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/28/2022

Efficient-Adam: Communication-Efficient Distributed Adam with Complexity Analysis

Distributed adaptive stochastic gradient methods have been widely used f...
research
11/12/2020

Distributed Sparse SGD with Majority Voting

Distributed learning, particularly variants of distributed stochastic gr...
research
11/03/2017

PS-DBSCAN: An Efficient Parallel DBSCAN Algorithm Based on Platform Of AI (PAI)

We present PS-DBSCAN, a communication efficient parallel DBSCAN algorith...
research
12/02/2020

Ship Detection: Parameter Server Variant

Deep learning ship detection in satellite optical imagery suffers from f...
research
10/11/2018

signSGD with Majority Vote is Communication Efficient And Byzantine Fault Tolerant

Training neural networks on large datasets can be accelerated by distrib...
research
04/22/2018

MQGrad: Reinforcement Learning of Gradient Quantization in Parameter Server

One of the most significant bottleneck in training large scale machine l...
research
04/03/2023

SparDL: Distributed Deep Learning Training with Efficient Sparse Communication

Top-k sparsification has recently been widely used to reduce the communi...

Please sign up or login with your details

Forgot password? Click here to reset