Sparse-SignSGD with Majority Vote for Communication-Efficient Distributed Learning
The training efficiency of complex deep learning models can be significantly improved through the use of distributed optimization. However, this process is often hindered by a large amount of communication cost between workers and a parameter server during iterations. To address this bottleneck, in this paper, we present a new communication-efficient algorithm that offers the synergistic benefits of both sparsification and sign quantization, called S^3GD-MV. The workers in S^3GD-MV select the top-K magnitude components of their local gradient vector and only send the signs of these components to the server. The server then aggregates the signs and returns the results via a majority vote rule. Our analysis shows that, under certain mild conditions, S^3GD-MV can converge at the same rate as signSGD while significantly reducing communication costs, if the sparsification parameter K is properly chosen based on the number of workers and the size of the deep learning model. Experimental results using both independent and identically distributed (IID) and non-IID datasets demonstrate that the S^3GD-MV attains higher accuracy than signSGD, significantly reducing communication costs. These findings highlight the potential of S^3GD-MV as a promising solution for communication-efficient distributed optimization in deep learning.
READ FULL TEXT