CEDAS: A Compressed Decentralized Stochastic Gradient Method with Improved Convergence
In this paper, we consider solving the distributed optimization problem over a multi-agent network under the communication restricted setting. We study a compressed decentralized stochastic gradient method, termed “compressed exact diffusion with adaptive stepsizes (CEDAS)", and show the method asymptotically achieves comparable convergence rate as centralized SGD for both smooth strongly convex objective functions and smooth nonconvex objective functions under unbiased compression operators. In particular, to our knowledge, CEDAS enjoys so far the shortest transient time (with respect to the graph specifics) for achieving the convergence rate of centralized SGD, which behaves as 𝒪(nC^3/(1-λ_2)^2) under smooth strongly convex objective functions, and 𝒪(n^3C^6/(1-λ_2)^4) under smooth nonconvex objective functions, where (1-λ_2) denotes the spectral gap of the mixing matrix, and C>0 is the compression-related parameter. Numerical experiments further demonstrate the effectiveness of the proposed algorithm.
READ FULL TEXT