DADAO: Decoupled Accelerated Decentralized Asynchronous Optimization for Time-Varying Gossips
DADAO is a novel decentralized asynchronous stochastic algorithm to minimize a sum of L-smooth and μ-strongly convex functions distributed over a time-varying connectivity network of size n. We model the local gradient updates and gossip communication procedures with separate independent Poisson Point Processes, decoupling the computation and communication steps in addition to making the whole approach completely asynchronous. Our method employs primal gradients and do not use a multi-consensus inner loop nor other ad-hoc mechanisms as Error Feedback, Gradient Tracking or a Proximal operator. By relating spatial quantities of our graphs χ^*_1,χ_2^* to a necessary minimal communication rate between nodes of the network, we show that our algorithm requires 𝒪(n√(L/μ)logϵ) local gradients and only 𝒪(n√(χ_1^*χ_2^*)√(L/μ)logϵ) communications to reach a precision ϵ. If SGD with uniform noise σ^2 is used, we reach a precision ϵ with same speed, up to a bias term in 𝒪(σ^2/√(μ L)). This improves upon the bounds obtained with current state-of-the-art approaches, our simulations validating the strength of our relatively unconstrained method. Our source-code is released on a public repository.
READ FULL TEXT