The Generative Leap: Sharp Sample Complexity for Efficiently Learning Gaussian Multi-Index Models

In this work we consider generic Gaussian Multi-index models, in which the labels only depend on the (Gaussian) -dimensional inputs through their projection onto a low-dimensional subspace, and we study efficient agnostic estimation procedures for this hidden subspace. We introduce the \emph{generative leap} exponent , a natural extension of the generative exponent from [Damian et al.'24] to the multi-index setting. We first show that a sample complexity of is necessary in the class of algorithms captured by the Low-Degree-Polynomial framework. We then establish that this sample complexity is also sufficient, by giving an agnostic sequential estimation procedure (that is, requiring no prior knowledge of the multi-index model) based on a spectral U-statistic over appropriate Hermite tensors. We further compute the generative leap exponent for several examples including piecewise linear functions (deep ReLU networks with bias), and general deep neural networks (with -dimensional first hidden layer).
View on arXiv@article{damian2025_2506.05500, title={ The Generative Leap: Sharp Sample Complexity for Efficiently Learning Gaussian Multi-Index Models }, author={ Alex Damian and Jason D. Lee and Joan Bruna }, journal={arXiv preprint arXiv:2506.05500}, year={ 2025 } }