56
8

A Note on Connectivity of Sublevel Sets in Deep Learning

Abstract

It is shown that for deep neural networks, a single wide layer of width N+1N+1 (NN being the number of training samples) suffices to prove the connectivity of sublevel sets of the training loss function. In the two-layer setting, the same property may not hold even if one has just one neuron less (i.e. width NN can lead to disconnected sublevel sets).

View on arXiv
Comments on this paper