ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2210.09371
21
7

On Accelerated Perceptrons and Beyond

17 October 2022
Guanghui Wang
Rafael Hanashiro
E. Guha
Jacob D. Abernethy
ArXivPDFHTML
Abstract

The classical Perceptron algorithm of Rosenblatt can be used to find a linear threshold function to correctly classify nnn linearly separable data points, assuming the classes are separated by some margin γ>0\gamma > 0γ>0. A foundational result is that Perceptron converges after Ω(1/γ2)\Omega(1/\gamma^{2})Ω(1/γ2) iterations. There have been several recent works that managed to improve this rate by a quadratic factor, to Ω(log⁡n/γ)\Omega(\sqrt{\log n}/\gamma)Ω(logn​/γ), with more sophisticated algorithms. In this paper, we unify these existing results under one framework by showing that they can all be described through the lens of solving min-max problems using modern acceleration techniques, mainly through optimistic online learning. We then show that the proposed framework also lead to improved results for a series of problems beyond the standard Perceptron setting. Specifically, a) For the margin maximization problem, we improve the state-of-the-art result from O(log⁡t/t2)O(\log t/t^2)O(logt/t2) to O(1/t2)O(1/t^2)O(1/t2), where ttt is the number of iterations; b) We provide the first result on identifying the implicit bias property of the classical Nesterov's accelerated gradient descent (NAG) algorithm, and show NAG can maximize the margin with an O(1/t2)O(1/t^2)O(1/t2) rate; c) For the classical ppp-norm Perceptron problem, we provide an algorithm with Ω((p−1)log⁡n/γ)\Omega(\sqrt{(p-1)\log n}/\gamma)Ω((p−1)logn​/γ) convergence rate, while existing algorithms suffer the Ω((p−1)/γ2)\Omega({(p-1)}/\gamma^2)Ω((p−1)/γ2) convergence rate.

View on arXiv
Comments on this paper