TY - JOUR
T1 - Near-Optimal Algorithms for Minimax Optimization
AU - Lin, Tianyi
AU - Jin, Chi
AU - Jordan, Michael I.
N1 - Funding Information:
The authors acknowledge Guangzeng Xie for helpful comments on the proof of Theorem 6, 8 and 9. This work was supported in part by the Mathematical Data Science program of the Office of Naval Research under grant number N00014-18-1-2764.
Publisher Copyright:
© 2020 T. Lin, C. Jin & M.I. Jordan.
PY - 2020
Y1 - 2020
N2 - This paper resolves a longstanding open question pertaining to the design of near-optimal first-order algorithms for smooth and strongly-convex-strongly-concave minimax problems. Current state-of-the-art first-order algorithms find an approximate Nash equilibrium using Õ(κx+κy) (Tseng, 1995) or Õ(min{κx √κy, √κxκy}) (Alkousa et al., 2019) gradient evaluations, where κx and κy are the condition numbers for the strong-convexity and strong-concavity assumptions. A gap still remains between these results and the best existing lower bound Ω̃(√κxκy) (Ibrahim et al., 2019; Zhang et al., 2019). This paper presents the first algorithm with Õ(√κxκy) gradient complexity, matching the lower bound up to logarithmic factors. Our algorithm is designed based on an accelerated proximal point method and an accelerated solver for minimax proximal steps. It can be easily extended to the settings of strongly-convex-concave, convex-concave, nonconvex-strongly-concave, and nonconvex-concave functions. This paper also presents algorithms that match or outperform all existing methods in these settings in terms of gradient complexity, up to logarithmic factors.
AB - This paper resolves a longstanding open question pertaining to the design of near-optimal first-order algorithms for smooth and strongly-convex-strongly-concave minimax problems. Current state-of-the-art first-order algorithms find an approximate Nash equilibrium using Õ(κx+κy) (Tseng, 1995) or Õ(min{κx √κy, √κxκy}) (Alkousa et al., 2019) gradient evaluations, where κx and κy are the condition numbers for the strong-convexity and strong-concavity assumptions. A gap still remains between these results and the best existing lower bound Ω̃(√κxκy) (Ibrahim et al., 2019; Zhang et al., 2019). This paper presents the first algorithm with Õ(√κxκy) gradient complexity, matching the lower bound up to logarithmic factors. Our algorithm is designed based on an accelerated proximal point method and an accelerated solver for minimax proximal steps. It can be easily extended to the settings of strongly-convex-concave, convex-concave, nonconvex-strongly-concave, and nonconvex-concave functions. This paper also presents algorithms that match or outperform all existing methods in these settings in terms of gradient complexity, up to logarithmic factors.
UR - http://www.scopus.com/inward/record.url?scp=85135585065&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85135585065&partnerID=8YFLogxK
M3 - Conference article
AN - SCOPUS:85135585065
SN - 2640-3498
VL - 125
SP - 2738
EP - 2779
JO - Proceedings of Machine Learning Research
JF - Proceedings of Machine Learning Research
T2 - 33rd Conference on Learning Theory, COLT 2020
Y2 - 9 July 2020 through 12 July 2020
ER -