ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1811.03962
  4. Cited By
A Convergence Theory for Deep Learning via Over-Parameterization
v1v2v3v4v5 (latest)

A Convergence Theory for Deep Learning via Over-Parameterization

9 November 2018
Zeyuan Allen-Zhu
Yuanzhi Li
Zhao Song
    AI4CEODL
ArXiv (abs)PDFHTML

Papers citing "A Convergence Theory for Deep Learning via Over-Parameterization"

28 / 78 papers shown
Title
Recovery Guarantees for One-hidden-layer Neural Networks
Recovery Guarantees for One-hidden-layer Neural Networks
Kai Zhong
Zhao Song
Prateek Jain
Peter L. Bartlett
Inderjit S. Dhillon
MLT
175
337
0
10 Jun 2017
Convergence Analysis of Two-layer Neural Networks with ReLU Activation
Convergence Analysis of Two-layer Neural Networks with ReLU Activation
Yuanzhi Li
Yang Yuan
MLT
150
652
0
28 May 2017
Global Guarantees for Enforcing Deep Generative Priors by Empirical Risk
Global Guarantees for Enforcing Deep Generative Priors by Empirical Risk
Paul Hand
V. Voroninski
UQCV
138
138
0
22 May 2017
Learning ReLUs via Gradient Descent
Learning ReLUs via Gradient Descent
Mahdi Soltanolkotabi
MLT
75
182
0
10 May 2017
An Analytical Formula of Population Gradient for two-layered ReLU
  network and its Applications in Convergence and Critical Point Analysis
An Analytical Formula of Population Gradient for two-layered ReLU network and its Applications in Convergence and Critical Point Analysis
Yuandong Tian
MLT
189
216
0
02 Mar 2017
SGD Learns the Conjugate Kernel Class of the Network
SGD Learns the Conjugate Kernel Class of the Network
Amit Daniely
186
182
0
27 Feb 2017
Globally Optimal Gradient Descent for a ConvNet with Gaussian Inputs
Globally Optimal Gradient Descent for a ConvNet with Gaussian Inputs
Alon Brutzkus
Amir Globerson
MLT
168
313
0
26 Feb 2017
Convergence Results for Neural Networks via Electrodynamics
Convergence Results for Neural Networks via Electrodynamics
Rina Panigrahy
Sushant Sachdeva
Qiuyi Zhang
MLTMDE
77
22
0
01 Feb 2017
Follow the Compressed Leader: Faster Online Learning of Eigenvectors and
  Faster MMWU
Follow the Compressed Leader: Faster Online Learning of Eigenvectors and Faster MMWU
Zeyuan Allen-Zhu
Yuanzhi Li
66
44
0
06 Jan 2017
Reliably Learning the ReLU in Polynomial Time
Reliably Learning the ReLU in Polynomial Time
Surbhi Goel
Varun Kanade
Adam R. Klivans
J. Thaler
78
127
0
30 Nov 2016
Identity Matters in Deep Learning
Identity Matters in Deep Learning
Moritz Hardt
Tengyu Ma
OOD
84
398
0
14 Nov 2016
Understanding deep learning requires rethinking generalization
Understanding deep learning requires rethinking generalization
Chiyuan Zhang
Samy Bengio
Moritz Hardt
Benjamin Recht
Oriol Vinyals
HAI
339
4,629
0
10 Nov 2016
Gradient Descent Learns Linear Dynamical Systems
Gradient Descent Learns Linear Dynamical Systems
Moritz Hardt
Tengyu Ma
Benjamin Recht
107
240
0
16 Sep 2016
No bad local minima: Data independent training error guarantees for
  multilayer neural networks
No bad local minima: Data independent training error guarantees for multilayer neural networks
Daniel Soudry
Y. Carmon
185
235
0
26 May 2016
Wide Residual Networks
Wide Residual Networks
Sergey Zagoruyko
N. Komodakis
340
7,985
0
23 May 2016
Deep Learning without Poor Local Minima
Deep Learning without Poor Local Minima
Kenji Kawaguchi
ODL
219
923
0
23 May 2016
Deep Residual Learning for Image Recognition
Deep Residual Learning for Image Recognition
Kaiming He
Xinming Zhang
Shaoqing Ren
Jian Sun
MedIm
2.2K
194,020
0
10 Dec 2015
Deep Speech 2: End-to-End Speech Recognition in English and Mandarin
Deep Speech 2: End-to-End Speech Recognition in English and Mandarin
Dario Amodei
Rishita Anubhai
Eric Battenberg
Carl Case
Jared Casper
...
Chong-Jun Wang
Bo Xiao
Dani Yogatama
J. Zhan
Zhenyao Zhu
134
2,973
0
08 Dec 2015
Continuous control with deep reinforcement learning
Continuous control with deep reinforcement learning
Timothy Lillicrap
Jonathan J. Hunt
Alexander Pritzel
N. Heess
Tom Erez
Yuval Tassa
David Silver
Daan Wierstra
320
13,248
0
09 Sep 2015
Training Very Deep Networks
Training Very Deep Networks
R. Srivastava
Klaus Greff
Jürgen Schmidhuber
161
1,682
0
22 Jul 2015
Complexity Theoretic Limitations on Learning Halfspaces
Complexity Theoretic Limitations on Learning Halfspaces
Amit Daniely
103
141
0
21 May 2015
Qualitatively characterizing neural network optimization problems
Qualitatively characterizing neural network optimization problems
Ian Goodfellow
Oriol Vinyals
Andrew M. Saxe
ODL
110
522
0
19 Dec 2014
On the Computational Efficiency of Training Neural Networks
On the Computational Efficiency of Training Neural Networks
Roi Livni
Shai Shalev-Shwartz
Ohad Shamir
143
480
0
05 Oct 2014
Going Deeper with Convolutions
Going Deeper with Convolutions
Christian Szegedy
Wei Liu
Yangqing Jia
P. Sermanet
Scott E. Reed
Dragomir Anguelov
D. Erhan
Vincent Vanhoucke
Andrew Rabinovich
468
43,658
0
17 Sep 2014
Very Deep Convolutional Networks for Large-Scale Image Recognition
Very Deep Convolutional Networks for Large-Scale Image Recognition
Karen Simonyan
Andrew Zisserman
FAttMDE
1.6K
100,386
0
04 Sep 2014
Complexity theoretic limitations on learning DNF's
Complexity theoretic limitations on learning DNF's
Amit Daniely
Shai Shalev-Shwartz
59
112
0
13 Apr 2014
Speech Recognition with Deep Recurrent Neural Networks
Speech Recognition with Deep Recurrent Neural Networks
Alex Graves
Abdel-rahman Mohamed
Geoffrey E. Hinton
226
8,517
0
22 Mar 2013
A Variant of Azuma's Inequality for Martingales with Subgaussian Tails
A Variant of Azuma's Inequality for Martingales with Subgaussian Tails
Ohad Shamir
74
37
0
11 Oct 2011
Previous
12