ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1303.6149
63
164

Adaptivity of averaged stochastic gradient descent to local strong convexity for logistic regression

25 March 2013
Francis R. Bach
ArXivPDFHTML
Abstract

In this paper, we consider supervised learning problems such as logistic regression and study the stochastic gradient method with averaging, in the usual stochastic approximation setting where observations are used only once. We show that after NNN iterations, with a constant step-size proportional to 1/R2N1/R^2 \sqrt{N}1/R2N​ where NNN is the number of observations and RRR is the maximum norm of the observations, the convergence rate is always of order O(1/N)O(1/\sqrt{N})O(1/N​), and improves to O(R2/μN)O(R^2 / \mu N)O(R2/μN) where μ\muμ is the lowest eigenvalue of the Hessian at the global optimum (when this eigenvalue is greater than R2/NR^2/\sqrt{N}R2/N​). Since μ\muμ does not need to be known in advance, this shows that averaged stochastic gradient is adaptive to \emph{unknown local} strong convexity of the objective function. Our proof relies on the generalized self-concordance properties of the logistic loss and thus extends to all generalized linear models with uniformly bounded features.

View on arXiv
Comments on this paper