ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1810.06793
28
57

Learning Two-layer Neural Networks with Symmetric Inputs

16 October 2018
Rong Ge
Rohith Kuditipudi
Zhize Li
Xiang Wang
    OOD
    MLT
ArXivPDFHTML
Abstract

We give a new algorithm for learning a two-layer neural network under a general class of input distributions. Assuming there is a ground-truth two-layer network y = A \sigma(Wx) + \xi, where A,WA,WA,W are weight matrices, ξ\xiξ represents noise, and the number of neurons in the hidden layer is no larger than the input or output, our algorithm is guaranteed to recover the parameters A,WA,WA,W of the ground-truth network. The only requirement on the input xxx is that it is symmetric, which still allows highly complicated and structured input. Our algorithm is based on the method-of-moments framework and extends several results in tensor decompositions. We use spectral algorithms to avoid the complicated non-convex optimization in learning neural networks. Experiments show that our algorithm can robustly learn the ground-truth neural network with a small number of samples for many symmetric input distributions.

View on arXiv
Comments on this paper