20
3

Activation Functions Not To Active: A Plausible Theory on Interpreting Neural Networks

Abstract

Researchers commonly believe that neural networks model a high-dimensional space but cannot give a clear definition of this space. What is this space? What is its dimension? And does it has finite dimensions? In this paper, we develop a plausible theory on interpreting neural networks in terms of the role of activation functions in neural networks and define a high-dimensional (more precisely, an infinite-dimensional) space that neural networks including deep-learning networks could create. We show that the activation function acts as a magnifying function that maps the low-dimensional linear space into an infinite-dimensional space, which can distinctly identify the polynomial approximation of any multivariate continuous function of the variable values being the same features of the given dataset. Given a dataset with each example of dd features f1f_1, f2f_2, \cdots, fdf_d, we believe that neural networks model a special space with infinite dimensions, each of which is a monomial \prod_{i_1, i_2, \cdots, i_d} f_1^{i_1} f_2^{i_2} \cdots f_d^{i_d} for some non-negative integers i1,i2,,idZ0+={0,1,2,3,}{i_1, i_2, \cdots, i_d} \in \mathbb{Z}_{0}^{+}=\{0,1,2,3,\ldots\} . We term such an infinite-dimensional space a  Super Space (SS)\textit{ Super Space (SS)}. We see such a dimension as the minimum information unit. Every neuron node previously through an activation layer in neural networks is a  Super Plane (SP) \textit{ Super Plane (SP) }, which is actually a polynomial of infinite degree. This  Super Space \textit{ Super Space } is something like a coordinate system, in which every multivalue function can be represented by a  Super Plane \textit{ Super Plane }. We also show that training NNs could at least be reduced to solving a system of nonlinear equations. %solve sets of nonlinear equations

View on arXiv
Comments on this paper