271

Stochastic Conditional Gradient++

Abstract

In this paper, we develop Stochastic Continuous Greedy++ (SCG++), the first efficient variant of a conditional gradient method for maximizing a continuous submodular function subject to a convex constraint. Concretely, for a monotone and continuous DR-submodular function, SCG++ achieves a tight [(11/e)OPTϵ][(1-1/e)\text{OPT} -\epsilon] solution while using O(1/ϵ2)O(1/\epsilon^2) stochastic oracle queries and O(1/ϵ)O(1/\epsilon) calls to the linear optimization oracle. The best previously known algorithms either achieve a suboptimal [(1/2)OPTϵ][(1/2)\text{OPT} -\epsilon] solution with O(1/ϵ2)O(1/\epsilon^2) stochastic gradients or the tight [(11/e)OPTϵ][(1-1/e)\text{OPT} -\epsilon] solution with suboptimal O(1/ϵ3)O(1/\epsilon^3) stochastic gradients. SCG++ enjoys optimality in terms of both approximation guarantee and stochastic stochastic oracle queries. Our novel variance reduction method naturally extends to stochastic convex minimization. More precisely, we develop Stochastic Frank-Wolfe++ (SFW++) that achieves an ϵ\epsilon-approximate optimum with only O(1/ϵ)O(1/\epsilon) calls to the linear optimization oracle while using O(1/ϵ2)O(1/\epsilon^2) stochastic oracle queries in total. Therefore, SFW++ is the first efficient projection-free algorithm that achieves the optimum complexity O(1/ϵ2)O(1/\epsilon^2) in terms of stochastic oracle queries.

View on arXiv
Comments on this paper