334

Finite-Sum Optimization: A New Perspective for Convergence to a Global Solution

Main:27 Pages
Bibliography:1 Pages
1 Tables
Appendix:1 Pages
Abstract

Deep neural networks (DNNs) have shown great success in many machine learning tasks. Their training is challenging since the loss surface of the network architecture is generally non-convex, or even non-smooth. How and under what assumptions is guaranteed convergence to a \textit{global} minimum possible? We propose a reformulation of the minimization problem allowing for a new recursive algorithmic framework. By using bounded style assumptions, we prove convergence to an ε\varepsilon-(global) minimum using O~(1/ε3)\mathcal{\tilde{O}}(1/\varepsilon^3) gradient computations. Our theoretical foundation motivates further study, implementation, and optimization of the new algorithmic framework and further investigation of its non-standard bounded style assumptions. This new direction broadens our understanding of why and under what circumstances training of a DNN converges to a global minimum.

View on arXiv
Comments on this paper