30
1

The Generative Leap: Sharp Sample Complexity for Efficiently Learning Gaussian Multi-Index Models

Main:14 Pages
Bibliography:5 Pages
Appendix:19 Pages
Abstract

In this work we consider generic Gaussian Multi-index models, in which the labels only depend on the (Gaussian) dd-dimensional inputs through their projection onto a low-dimensional r=Od(1)r = O_d(1) subspace, and we study efficient agnostic estimation procedures for this hidden subspace. We introduce the \emph{generative leap} exponent kk^\star, a natural extension of the generative exponent from [Damian et al.'24] to the multi-index setting. We first show that a sample complexity of n=Θ(d1\k/2)n=\Theta(d^{1 \vee \k/2}) is necessary in the class of algorithms captured by the Low-Degree-Polynomial framework. We then establish that this sample complexity is also sufficient, by giving an agnostic sequential estimation procedure (that is, requiring no prior knowledge of the multi-index model) based on a spectral U-statistic over appropriate Hermite tensors. We further compute the generative leap exponent for several examples including piecewise linear functions (deep ReLU networks with bias), and general deep neural networks (with rr-dimensional first hidden layer).

View on arXiv
@article{damian2025_2506.05500,
  title={ The Generative Leap: Sharp Sample Complexity for Efficiently Learning Gaussian Multi-Index Models },
  author={ Alex Damian and Jason D. Lee and Joan Bruna },
  journal={arXiv preprint arXiv:2506.05500},
  year={ 2025 }
}
Comments on this paper