ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1605.07110
  4. Cited By
Deep Learning without Poor Local Minima

Deep Learning without Poor Local Minima

23 May 2016
Kenji Kawaguchi
    ODL
ArXivPDFHTML

Papers citing "Deep Learning without Poor Local Minima"

50 / 207 papers shown
Title
Defending Against Saddle Point Attack in Byzantine-Robust Distributed
  Learning
Defending Against Saddle Point Attack in Byzantine-Robust Distributed Learning
Dong Yin
Yudong Chen
Kannan Ramchandran
Peter L. Bartlett
FedML
32
97
0
14 Jun 2018
Representation Learning on Graphs with Jumping Knowledge Networks
Representation Learning on Graphs with Jumping Knowledge Networks
Keyulu Xu
Chengtao Li
Yonglong Tian
Tomohiro Sonobe
Ken-ichi Kawarabayashi
Stefanie Jegelka
GNN
279
1,944
0
09 Jun 2018
Towards Riemannian Accelerated Gradient Methods
Towards Riemannian Accelerated Gradient Methods
Hongyi Zhang
S. Sra
19
53
0
07 Jun 2018
Deep Neural Networks with Multi-Branch Architectures Are Less Non-Convex
Deep Neural Networks with Multi-Branch Architectures Are Less Non-Convex
Hongyang R. Zhang
Junru Shao
Ruslan Salakhutdinov
39
14
0
06 Jun 2018
Challenges in High-dimensional Reinforcement Learning with Evolution
  Strategies
Challenges in High-dimensional Reinforcement Learning with Evolution Strategies
Nils Müller
Tobias Glasmachers
33
28
0
04 Jun 2018
Stochastic Gradient/Mirror Descent: Minimax Optimality and Implicit
  Regularization
Stochastic Gradient/Mirror Descent: Minimax Optimality and Implicit Regularization
Navid Azizan
B. Hassibi
24
61
0
04 Jun 2018
Algorithmic Regularization in Learning Deep Homogeneous Models: Layers
  are Automatically Balanced
Algorithmic Regularization in Learning Deep Homogeneous Models: Layers are Automatically Balanced
S. Du
Wei Hu
J. Lee
MLT
40
237
0
04 Jun 2018
Understanding Generalization and Optimization Performance of Deep CNNs
Understanding Generalization and Optimization Performance of Deep CNNs
Pan Zhou
Jiashi Feng
MLT
30
48
0
28 May 2018
Adding One Neuron Can Eliminate All Bad Local Minima
Adding One Neuron Can Eliminate All Bad Local Minima
Shiyu Liang
Ruoyu Sun
J. Lee
R. Srikant
39
89
0
22 May 2018
How Many Samples are Needed to Estimate a Convolutional or Recurrent
  Neural Network?
How Many Samples are Needed to Estimate a Convolutional or Recurrent Neural Network?
S. Du
Yining Wang
Xiyu Zhai
Sivaraman Balakrishnan
Ruslan Salakhutdinov
Aarti Singh
SSL
21
57
0
21 May 2018
Improved Learning of One-hidden-layer Convolutional Neural Networks with
  Overlaps
Improved Learning of One-hidden-layer Convolutional Neural Networks with Overlaps
S. Du
Surbhi Goel
MLT
30
17
0
20 May 2018
The Global Optimization Geometry of Shallow Linear Neural Networks
The Global Optimization Geometry of Shallow Linear Neural Networks
Zhihui Zhu
Daniel Soudry
Yonina C. Eldar
M. Wakin
ODL
20
36
0
13 May 2018
The Loss Surface of XOR Artificial Neural Networks
The Loss Surface of XOR Artificial Neural Networks
D. Mehta
Xiaojun Zhao
Edgar A. Bernal
D. Wales
34
19
0
06 Apr 2018
Policy Search in Continuous Action Domains: an Overview
Policy Search in Continuous Action Domains: an Overview
Olivier Sigaud
F. Stulp
16
72
0
13 Mar 2018
How to Start Training: The Effect of Initialization and Architecture
How to Start Training: The Effect of Initialization and Architecture
Boris Hanin
David Rolnick
13
253
0
05 Mar 2018
Essentially No Barriers in Neural Network Energy Landscape
Essentially No Barriers in Neural Network Energy Landscape
Felix Dräxler
K. Veschgini
M. Salmhofer
Fred Hamprecht
MoMe
22
424
0
02 Mar 2018
Understanding the Loss Surface of Neural Networks for Binary
  Classification
Understanding the Loss Surface of Neural Networks for Binary Classification
Shiyu Liang
Ruoyu Sun
Yixuan Li
R. Srikant
35
87
0
19 Feb 2018
Gradient descent with identity initialization efficiently learns
  positive definite linear transformations by deep residual networks
Gradient descent with identity initialization efficiently learns positive definite linear transformations by deep residual networks
Peter L. Bartlett
D. Helmbold
Philip M. Long
36
116
0
16 Feb 2018
Deep Neural Networks Learn Non-Smooth Functions Effectively
Deep Neural Networks Learn Non-Smooth Functions Effectively
Masaaki Imaizumi
Kenji Fukumizu
23
123
0
13 Feb 2018
Visualizing the Loss Landscape of Neural Nets
Visualizing the Loss Landscape of Neural Nets
Hao Li
Zheng Xu
Gavin Taylor
Christoph Studer
Tom Goldstein
108
1,848
0
28 Dec 2017
Accelerated Gradient Descent Escapes Saddle Points Faster than Gradient
  Descent
Accelerated Gradient Descent Escapes Saddle Points Faster than Gradient Descent
Chi Jin
Praneeth Netrapalli
Michael I. Jordan
ODL
37
261
0
28 Nov 2017
High-dimensional dynamics of generalization error in neural networks
High-dimensional dynamics of generalization error in neural networks
Madhu S. Advani
Andrew M. Saxe
AI4CE
90
464
0
10 Oct 2017
AdaDNNs: Adaptive Ensemble of Deep Neural Networks for Scene Text
  Recognition
AdaDNNs: Adaptive Ensemble of Deep Neural Networks for Scene Text Recognition
Chun Yang
Xu-Cheng Yin
Zejun Li
Jianwei Wu
Chunchao Guo
Hongfa Wang
Lei Xiao
24
10
0
10 Oct 2017
Mini-batch Tempered MCMC
Mini-batch Tempered MCMC
Dangna Li
W. Wong
26
5
0
31 Jul 2017
Cosmological model discrimination with Deep Learning
Cosmological model discrimination with Deep Learning
Jorit Schmelzle
Aurelien Lucchi
T. Kacprzak
A. Amara
R. Sgier
Alexandre Réfrégier
Thomas Hofmann
31
38
0
17 Jul 2017
Theoretical insights into the optimization landscape of
  over-parameterized shallow neural networks
Theoretical insights into the optimization landscape of over-parameterized shallow neural networks
Mahdi Soltanolkotabi
Adel Javanmard
J. Lee
36
415
0
16 Jul 2017
Global optimality conditions for deep neural networks
Global optimality conditions for deep neural networks
Chulhee Yun
S. Sra
Ali Jadbabaie
128
117
0
08 Jul 2017
Recovery Guarantees for One-hidden-layer Neural Networks
Recovery Guarantees for One-hidden-layer Neural Networks
Kai Zhong
Zhao Song
Prateek Jain
Peter L. Bartlett
Inderjit S. Dhillon
MLT
34
336
0
10 Jun 2017
Are Saddles Good Enough for Deep Learning?
Are Saddles Good Enough for Deep Learning?
Adepu Ravi Sankar
V. Balasubramanian
43
5
0
07 Jun 2017
On the stable recovery of deep structured linear networks under sparsity
  constraints
On the stable recovery of deep structured linear networks under sparsity constraints
F. Malgouyres
32
7
0
31 May 2017
Spectral Norm Regularization for Improving the Generalizability of Deep
  Learning
Spectral Norm Regularization for Improving the Generalizability of Deep Learning
Yuichi Yoshida
Takeru Miyato
40
325
0
31 May 2017
Deep Learning is Robust to Massive Label Noise
Deep Learning is Robust to Massive Label Noise
David Rolnick
Andreas Veit
Serge J. Belongie
Nir Shavit
NoLa
36
549
0
30 May 2017
The loss surface of deep and wide neural networks
The loss surface of deep and wide neural networks
Quynh N. Nguyen
Matthias Hein
ODL
51
283
0
26 Apr 2017
Snapshot Ensembles: Train 1, get M for free
Snapshot Ensembles: Train 1, get M for free
Gao Huang
Yixuan Li
Geoff Pleiss
Zhuang Liu
J. Hopcroft
Kilian Q. Weinberger
OOD
FedML
UQCV
47
935
0
01 Apr 2017
Langevin Dynamics with Continuous Tempering for Training Deep Neural
  Networks
Langevin Dynamics with Continuous Tempering for Training Deep Neural Networks
Nanyang Ye
Zhanxing Zhu
Rafał K. Mantiuk
19
20
0
13 Mar 2017
Evolution Strategies as a Scalable Alternative to Reinforcement Learning
Evolution Strategies as a Scalable Alternative to Reinforcement Learning
Tim Salimans
Jonathan Ho
Xi Chen
Szymon Sidor
Ilya Sutskever
33
1,516
0
10 Mar 2017
Data-Dependent Stability of Stochastic Gradient Descent
Data-Dependent Stability of Stochastic Gradient Descent
Ilja Kuzborskij
Christoph H. Lampert
MLT
9
165
0
05 Mar 2017
How to Escape Saddle Points Efficiently
How to Escape Saddle Points Efficiently
Chi Jin
Rong Ge
Praneeth Netrapalli
Sham Kakade
Michael I. Jordan
ODL
37
832
0
02 Mar 2017
Deep Semi-Random Features for Nonlinear Function Approximation
Deep Semi-Random Features for Nonlinear Function Approximation
Kenji Kawaguchi
Bo Xie
Vikas Verma
Le Song
29
15
0
28 Feb 2017
How ConvNets model Non-linear Transformations
How ConvNets model Non-linear Transformations
Dipan K. Pal
Marios Savvides
21
0
0
24 Feb 2017
Convergence Results for Neural Networks via Electrodynamics
Convergence Results for Neural Networks via Electrodynamics
Rina Panigrahy
Sushant Sachdeva
Qiuyi Zhang
MLT
MDE
29
22
0
01 Feb 2017
An empirical analysis of the optimization of deep network loss surfaces
An empirical analysis of the optimization of deep network loss surfaces
Daniel Jiwoong Im
Michael Tao
K. Branson
ODL
35
61
0
13 Dec 2016
Reliably Learning the ReLU in Polynomial Time
Reliably Learning the ReLU in Polynomial Time
Surbhi Goel
Varun Kanade
Adam R. Klivans
J. Thaler
29
124
0
30 Nov 2016
Geometric deep learning: going beyond Euclidean data
Geometric deep learning: going beyond Euclidean data
M. Bronstein
Joan Bruna
Yann LeCun
Arthur Szlam
P. Vandergheynst
GNN
264
3,243
0
24 Nov 2016
Local minima in training of neural networks
Local minima in training of neural networks
G. Swirszcz
Wojciech M. Czarnecki
Razvan Pascanu
ODL
37
73
0
19 Nov 2016
Identity Matters in Deep Learning
Identity Matters in Deep Learning
Moritz Hardt
Tengyu Ma
OOD
25
398
0
14 Nov 2016
Diverse Neural Network Learns True Target Functions
Diverse Neural Network Learns True Target Functions
Bo Xie
Yingyu Liang
Le Song
14
137
0
09 Nov 2016
Topology and Geometry of Half-Rectified Network Optimization
Topology and Geometry of Half-Rectified Network Optimization
C. Freeman
Joan Bruna
19
233
0
04 Nov 2016
Understanding Deep Neural Networks with Rectified Linear Units
Understanding Deep Neural Networks with Rectified Linear Units
R. Arora
A. Basu
Poorya Mianjy
Anirbit Mukherjee
PINN
41
634
0
04 Nov 2016
Demystifying ResNet
Demystifying ResNet
Sihan Li
Jiantao Jiao
Yanjun Han
Tsachy Weissman
32
38
0
03 Nov 2016
Previous
12345
Next