TY - GEN
T1 - On gradient descent ascent for nonconvex-concave minimax problems
AU - Lin, Tianyi
AU - Jin, Chi
AU - Jordan, Michael I.
N1 - Publisher Copyright:
© International Conference on Machine Learning, ICML 2020. All rights reserved.
PY - 2020
Y1 - 2020
N2 - We consider nonconvex-concave minimax problems, minxmaxy2Y f(x,y) where f is nonconvex in x but concave in y and Y is a convex and bounded set. One of the most popular algorithms for solving this problem is the celebrated gradient descent ascent (GDA) algorithm, which has been widely used in machine learning, control theory and economics. Despite the extensive convergence results for the convex-concave setting, GDA with equal stepsize can converge to limit cycles or even diverge in a general setting. In this paper, we present the complexity results on two-time-scale GDA for solving nonconvexconcave minimax problems, showing that the algorithm can find a stationary point of the function φ maxy2Y f( y) efficiently. To the best our knowledge, this is the first nonasymptotic analysis for two-time-scale GDA in this setting, shedding light on its superior practical performance in training generative adversarial networks (GANs) and other real applications.
AB - We consider nonconvex-concave minimax problems, minxmaxy2Y f(x,y) where f is nonconvex in x but concave in y and Y is a convex and bounded set. One of the most popular algorithms for solving this problem is the celebrated gradient descent ascent (GDA) algorithm, which has been widely used in machine learning, control theory and economics. Despite the extensive convergence results for the convex-concave setting, GDA with equal stepsize can converge to limit cycles or even diverge in a general setting. In this paper, we present the complexity results on two-time-scale GDA for solving nonconvexconcave minimax problems, showing that the algorithm can find a stationary point of the function φ maxy2Y f( y) efficiently. To the best our knowledge, this is the first nonasymptotic analysis for two-time-scale GDA in this setting, shedding light on its superior practical performance in training generative adversarial networks (GANs) and other real applications.
UR - http://www.scopus.com/inward/record.url?scp=85105571322&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85105571322&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85105571322
T3 - 37th International Conference on Machine Learning, ICML 2020
SP - 6039
EP - 6049
BT - 37th International Conference on Machine Learning, ICML 2020
A2 - Daume, Hal
A2 - Singh, Aarti
PB - International Machine Learning Society (IMLS)
T2 - 37th International Conference on Machine Learning, ICML 2020
Y2 - 13 July 2020 through 18 July 2020
ER -