A deep network construction that adapts to intrinsic dimensionality beyond the domain

We study the approximation of two-layer compositions via deep networks with ReLU activation, where is a geometrically intuitive, dimensionality reducing feature map. We focus on two intuitive and practically relevant choices for : the projection onto a low-dimensional embedded submanifold and a distance to a collection of low-dimensional sets. We achieve near optimal approximation rates, which depend only on the complexity of the dimensionality reducing map rather than the ambient dimension. Since encapsulates all nonlinear features that are material to the function , this suggests that deep nets are faithful to an intrinsic dimension governed by rather than the complexity of the domain of . In particular, the prevalent assumption of approximating functions on low-dimensional manifolds can be significantly relaxed using functions of type with representing an orthogonal projection onto the same manifold.
View on arXiv