Gradient-free stochastic optimization for additive models

We address the problem of zero-order optimization from noisy observations for an objective function satisfying the Polyak-Łojasiewicz or the strong convexity condition. Additionally, we assume that the objective function has an additive structure and satisfies a higher-order smoothness property, characterized by the Hölder family of functions. The additive model for Hölder classes of functions is well-studied in the literature on nonparametric function estimation, where it is shown that such a model benefits from a substantial improvement of the estimation accuracy compared to the Hölder model without additive structure. We study this established framework in the context of gradient-free optimization. We propose a randomized gradient estimator that, when plugged into a gradient descent algorithm, allows one to achieve minimax optimal optimization error of the order , where is the dimension of the problem, is the number of queries and is the Hölder degree of smoothness. We conclude that, in contrast to nonparametric estimation problems, no substantial gain of accuracy can be achieved when using additive models in gradient-free optimization.
View on arXiv@article{akhavan2025_2503.02131, title={ Gradient-free stochastic optimization for additive models }, author={ Arya Akhavan and Alexandre B. Tsybakov }, journal={arXiv preprint arXiv:2503.02131}, year={ 2025 } }