88
6

DoCoM: Compressed Decentralized Optimization with Near-Optimal Sample Complexity

Abstract

This paper proposes the Doubly Compressed Momentum-assisted stochastic gradient tracking algorithm DoCoM\texttt{DoCoM} for communication-efficient decentralized optimization. The algorithm features two main ingredients to achieve a near-optimal sample complexity while allowing for communication compression. First, the algorithm tracks both the averaged iterate and stochastic gradient using compressed gossiping consensus. Second, a momentum step is incorporated for adaptive variance reduction with the local gradient estimates. We show that DoCoM\texttt{DoCoM} finds a near-stationary solution at all participating agents satisfying E[f(θ)2]=O(1/T2/3)\mathbb{E}[ \| \nabla f( \theta ) \|^2 ] = \mathcal{O}( 1 / T^{2/3} ) in TT iterations, where f(θ)f(\theta) is a smooth (possibly non-convex) objective function. Notice that the proof is achieved via analytically designing a new potential function that tightly tracks the one-iteration progress of DoCoM\texttt{DoCoM}. As a corollary, our analysis also established the linear convergence of DoCoM\texttt{DoCoM} to a global optimal solution for objective functions with the Polyak-{\L}ojasiewicz condition. Numerical experiments demonstrate that our algorithm outperforms several state-of-the-art algorithms in practice.

View on arXiv
Comments on this paper