Smoothness, Low-Noise and Fast Rates

Abstract
We establish am excess risk bound of order for ERM with an H-smooth loss function and a hypothesis class with Rademacher complexity , where is the best risk achievable by the hypothesis class. For typical hypothesis classes where , this translates to a learning rate of order in the separable () case and more generally. We also provide similar guarantees for online and stochastic convex optimization of a smooth non-negative objective.
View on arXivComments on this paper