ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1511.07948
17
35

Learning Halfspaces and Neural Networks with Random Initialization

25 November 2015
Yuchen Zhang
J. Lee
Martin J. Wainwright
Michael I. Jordan
ArXivPDFHTML
Abstract

We study non-convex empirical risk minimization for learning halfspaces and neural networks. For loss functions that are LLL-Lipschitz continuous, we present algorithms to learn halfspaces and multi-layer neural networks that achieve arbitrarily small excess risk ϵ>0\epsilon>0ϵ>0. The time complexity is polynomial in the input dimension ddd and the sample size nnn, but exponential in the quantity (L/ϵ2)log⁡(L/ϵ)(L/\epsilon^2)\log(L/\epsilon)(L/ϵ2)log(L/ϵ). These algorithms run multiple rounds of random initialization followed by arbitrary optimization steps. We further show that if the data is separable by some neural network with constant margin γ>0\gamma>0γ>0, then there is a polynomial-time algorithm for learning a neural network that separates the training data with margin Ω(γ)\Omega(\gamma)Ω(γ). As a consequence, the algorithm achieves arbitrary generalization error ϵ>0\epsilon>0ϵ>0 with poly(d,1/ϵ){\rm poly}(d,1/\epsilon)poly(d,1/ϵ) sample and time complexity. We establish the same learnability result when the labels are randomly flipped with probability η<1/2\eta<1/2η<1/2.

View on arXiv
Comments on this paper