1.2K

On Gradient Descent Ascent for Nonconvex-Concave Minimax Problems

International Conference on Machine Learning (ICML), 2019
Abstract

We consider nonconvex-concave minimax problems, minxmaxyYf(x,y)\min_{x} \max_{y\in\mathcal{Y}} f(x, y), where ff is nonconvex in xx but concave in yy. The standard algorithm for solving this problem is the celebrated gradient descent ascent (GDA) algorithm, which has been widely used in machine learning, control theory and economics. However, despite the solid theory for the convex-concave setting, GDA can converge to limit cycles or even diverge in a general setting. In this paper, we present a nonasymptotic analysis of GDA for solving nonconvex-concave minimax problems, showing that GDA can find a stationary point of the function Φ():=maxyYf(,y)\Phi(\cdot) :=\max_{y\in\mathcal{Y} }f(\cdot, y) efficiently. To the best our knowledge, this is the first theoretical guarantee for GDA in this setting, shedding light on its practical performance in many real applications.

View on arXiv
Comments on this paper