Improve the Robustness and Accuracy of Deep Neural Network with Normalization

In this paper, the robustness and accuracy of the deep neural network (DNN) was enhanced by introducing the normalization of the weight matrices of the DNN with Relu as the activation function. It is proved that the normalization leads to large dihedral angles between two adjacent faces of the polyhedron graph of the DNN function and hence smoother DNN functions, which reduces over-fitting. A measure is proposed for the robustness of a classification DNN, which is the average radius of the maximal robust spheres with the sample data as centers. A lower bound for the robustness measure is given in terms of the norm. Finally, an upper bound for the Rademacher complexity of DNN with normalization is given. An algorithm is given to train a DNN with the normalization and experimental results are used to show that the normalization is effective to improve the robustness and accuracy.
View on arXiv