68
1

Convergence of Decentralized Stochastic Subgradient-based Methods for Nonsmooth Nonconvex functions

Abstract

In this paper, we focus on the decentralized stochastic subgradient-based methods in minimizing nonsmooth nonconvex functions without Clarke regularity, especially in the decentralized training of nonsmooth neural networks. We propose a general framework that unifies various decentralized subgradient-based methods, such as decentralized stochastic subgradient descent (DSGD), DSGD with gradient-tracking technique (DSGD-T), and DSGD with momentum (DSGD-M). To establish the convergence properties of our proposed framework, we relate the discrete iterates to the trajectories of a continuous-time differential inclusion, which is assumed to have a coercive Lyapunov function with a stable set A\mathcal{A}. We prove the asymptotic convergence of the iterates to the stable set A\mathcal{A} with sufficiently small and diminishing step-sizes. These results provide first convergence guarantees for some well-recognized of decentralized stochastic subgradient-based methods without Clarke regularity of the objective function. Preliminary numerical experiments demonstrate that our proposed framework yields highly efficient decentralized stochastic subgradient-based methods with convergence guarantees in the training of nonsmooth neural networks.

View on arXiv
@article{zhang2025_2403.11565,
  title={ Convergence of Decentralized Stochastic Subgradient-based Methods for Nonsmooth Nonconvex functions },
  author={ Siyuan Zhang and Nachuan Xiao and Xin Liu },
  journal={arXiv preprint arXiv:2403.11565},
  year={ 2025 }
}
Comments on this paper

We use cookies and other tracking technologies to improve your browsing experience on our website, to show you personalized content and targeted ads, to analyze our website traffic, and to understand where our visitors are coming from. See our policy.