ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1905.01502
11
5

Improved Classification Rates for Localized SVMs

4 May 2019
Ingrid Blaschzyk
Ingo Steinwart
ArXivPDFHTML
Abstract

Localized support vector machines solve SVMs on many spatially defined small chunks and one of their main characteristics besides the computational benefit compared to global SVMs is the freedom of choosing arbitrary kernel and regularization parameter on each cell. We take advantage of this observation to derive global learning rates for localized SVMs with Gaussian kernels and hinge loss. Under certain assumptions our rates outperform known classification rates for localized SVMs, for global SVMs, and other learning algorithms based on e.g., plug-in rules, trees, or DNNs. These rates are achieved under a set of margin conditions that describe the behavior of the data-generating distribution, where no assumption on the existence of a density is made. We observe that a margin condition that relates the distance to the decision boundary to the amount of noise is crucial to obtain rates. The statistical analysis relies on a careful analysis of the excess risk which includes a separation of the input space into a subset that is close to the decision boundary and into a subset that is sufficiently far away. Moreover, we show that our rates are obtained adaptively, that is, without knowing the parameters resulting from the margin conditions.

View on arXiv
Comments on this paper