ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1712.04104
45
41
v1v2v3 (latest)

Convergence Rates for Deterministic and Stochastic Subgradient Methods Without Lipschitz Continuity

12 December 2017
Benjamin Grimmer
ArXiv (abs)PDFHTML
Abstract

We generalize the classic convergence rate theory for subgradient methods to apply to non-Lipschitz functions via a new measure of steepness. For the deterministic projected subgradient method, we derive a global O(1/T)O(1/\sqrt{T})O(1/T​) convergence rate for any function with at most exponential growth. Our approach implies generalizations of the standard convergence rates for gradient descent on functions with Lipschitz or H\"older continuous gradients. Further, we show a O(1/T)O(1/\sqrt{T})O(1/T​) convergence rate for the stochastic projected subgradient method on functions with at most quadratic growth, which improves to O(1/T)O(1/T)O(1/T) under strong convexity.

View on arXiv
Comments on this paper