ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2208.14318
37
3
v1v2 (latest)

Convergence Rates of Training Deep Neural Networks via Alternating Minimization Methods

30 August 2022
Jintao Xu
Chenglong Bao
W. Xing
ArXiv (abs)PDFHTML
Abstract

Training deep neural networks (DNNs) is an important and challenging optimization problem in machine learning due to its non-convexity and non-separable structure. The alternating minimization (AM) approaches split the composition structure of DNNs and have drawn great interest in the deep learning and optimization communities. In this paper, we propose a unified framework for analyzing the convergence rate of AM-type network training methods. Our analysis are based on the jjj-step sufficient decrease conditions and the Kurdyka-Lojasiewicz (KL) property, which relaxes the requirement of designing descent algorithms. We show the detailed local convergence rate if the KL exponent θ\thetaθ varies in [0,1)[0,1)[0,1). Moreover, the local R-linear convergence is discussed under a stronger jjj-step sufficient decrease condition.

View on arXiv
Comments on this paper