26
0

Gradient-free stochastic optimization for additive models

Abstract

We address the problem of zero-order optimization from noisy observations for an objective function satisfying the Polyak-Łojasiewicz or the strong convexity condition. Additionally, we assume that the objective function has an additive structure and satisfies a higher-order smoothness property, characterized by the Hölder family of functions. The additive model for Hölder classes of functions is well-studied in the literature on nonparametric function estimation, where it is shown that such a model benefits from a substantial improvement of the estimation accuracy compared to the Hölder model without additive structure. We study this established framework in the context of gradient-free optimization. We propose a randomized gradient estimator that, when plugged into a gradient descent algorithm, allows one to achieve minimax optimal optimization error of the order dT(β1)/βdT^{-(\beta-1)/\beta}, where dd is the dimension of the problem, TT is the number of queries and β2\beta\ge 2 is the Hölder degree of smoothness. We conclude that, in contrast to nonparametric estimation problems, no substantial gain of accuracy can be achieved when using additive models in gradient-free optimization.

View on arXiv
@article{akhavan2025_2503.02131,
  title={ Gradient-free stochastic optimization for additive models },
  author={ Arya Akhavan and Alexandre B. Tsybakov },
  journal={arXiv preprint arXiv:2503.02131},
  year={ 2025 }
}
Comments on this paper