13
42

Accelerated Decentralized Optimization with Local Updates for Smooth and Strongly Convex Objectives

Abstract

In this paper, we study the problem of minimizing a sum of smooth and strongly convex functions split over the nodes of a network in a decentralized fashion. We propose the algorithm ESDACDESDACD, a decentralized accelerated algorithm that only requires local synchrony. Its rate depends on the condition number κ\kappa of the local functions as well as the network topology and delays. Under mild assumptions on the topology of the graph, ESDACDESDACD takes a time O((τmax+Δmax)κ/γln(ϵ1))O((\tau_{\max} + \Delta_{\max})\sqrt{{\kappa}/{\gamma}}\ln(\epsilon^{-1})) to reach a precision ϵ\epsilon where γ\gamma is the spectral gap of the graph, τmax\tau_{\max} the maximum communication delay and Δmax\Delta_{\max} the maximum computation time. Therefore, it matches the rate of SSDASSDA, which is optimal when τmax=Ω(Δmax)\tau_{\max} = \Omega\left(\Delta_{\max}\right). Applying ESDACDESDACD to quadratic local functions leads to an accelerated randomized gossip algorithm of rate O(θgossip/n)O( \sqrt{\theta_{\rm gossip}/n}) where θgossip\theta_{\rm gossip} is the rate of the standard randomized gossip. To the best of our knowledge, it is the first asynchronous gossip algorithm with a provably improved rate of convergence of the second moment of the error. We illustrate these results with experiments in idealized settings.

View on arXiv
Comments on this paper