Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1903.03488
Cited By
Is Deeper Better only when Shallow is Good?
8 March 2019
Eran Malach
Shai Shalev-Shwartz
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Is Deeper Better only when Shallow is Good?"
10 / 10 papers shown
Title
Provable Guarantees for Nonlinear Feature Learning in Three-Layer Neural Networks
Eshaan Nichani
Alexandru Damian
Jason D. Lee
MLT
38
13
0
11 May 2023
When Deep Learning Meets Polyhedral Theory: A Survey
Joey Huchette
Gonzalo Muñoz
Thiago Serra
Calvin Tsay
AI4CE
94
32
0
29 Apr 2023
Hidden Progress in Deep Learning: SGD Learns Parities Near the Computational Limit
Boaz Barak
Benjamin L. Edelman
Surbhi Goel
Sham Kakade
Eran Malach
Cyril Zhang
30
123
0
18 Jul 2022
An initial alignment between neural network and target is needed for gradient descent to learn
Emmanuel Abbe
Elisabetta Cornacchia
Jan Hązła
Christopher Marquis
24
16
0
25 Feb 2022
Random Feature Amplification: Feature Learning and Generalization in Neural Networks
Spencer Frei
Niladri S. Chatterji
Peter L. Bartlett
MLT
30
29
0
15 Feb 2022
Interplay between depth of neural networks and locality of target functions
Takashi Mori
Masakuni Ueda
22
0
0
28 Jan 2022
Fractal Structure and Generalization Properties of Stochastic Optimization Algorithms
A. Camuto
George Deligiannidis
Murat A. Erdogdu
Mert Gurbuzbalaban
Umut cSimcsekli
Lingjiong Zhu
30
29
0
09 Jun 2021
The Connection Between Approximation, Depth Separation and Learnability in Neural Networks
Eran Malach
Gilad Yehudai
Shai Shalev-Shwartz
Ohad Shamir
21
20
0
31 Jan 2021
Hausdorff Dimension, Heavy Tails, and Generalization in Neural Networks
Umut Simsekli
Ozan Sener
George Deligiannidis
Murat A. Erdogdu
44
55
0
16 Jun 2020
Benefits of depth in neural networks
Matus Telgarsky
142
602
0
14 Feb 2016
1