34
6

Memory-Query Tradeoffs for Randomized Convex Optimization

Abstract

We show that any randomized first-order algorithm which minimizes a dd-dimensional, 11-Lipschitz convex function over the unit ball must either use Ω(d2δ)\Omega(d^{2-\delta}) bits of memory or make Ω(d1+δ/6o(1))\Omega(d^{1+\delta/6-o(1)}) queries, for any constant δ(0,1)\delta\in (0,1) and when the precision ϵ\epsilon is quasipolynomially small in dd. Our result implies that cutting plane methods, which use O~(d2)\tilde{O}(d^2) bits of memory and O~(d)\tilde{O}(d) queries, are Pareto-optimal among randomized first-order algorithms, and quadratic memory is required to achieve optimal query complexity for convex optimization.

View on arXiv
Comments on this paper