TY - JOUR
T1 - Sampling can be faster than optimization
AU - Ma, Yi An
AU - Chen, Yuansi
AU - Jin, Chi
AU - Flammarion, Nicolas
AU - Jordan, Michael I.
N1 - Publisher Copyright:
© 2019 National Academy of Sciences. All rights reserved.
PY - 2019/10/15
Y1 - 2019/10/15
N2 - Optimization algorithms and Monte Carlo sampling algorithms have provided the computational foundations for the rapid growth in applications of statistical machine learning in recent years. There is, however, limited theoretical understanding of the relationships between these 2 kinds of methodology, and limited understanding of relative strengths and weaknesses. Moreover, existing results have been obtained primarily in the setting of convex functions (for optimization) and log-concave functions (for sampling). In this setting, where local properties determine global properties, optimization algorithms are unsurprisingly more efficient computationally than sampling algorithms. We instead examine a class of nonconvex objective functions that arise in mixture modeling and multistable systems. In this nonconvex setting, we find that the computational complexity of sampling algorithms scales linearly with the model dimension while that of optimization algorithms scales exponentially.
AB - Optimization algorithms and Monte Carlo sampling algorithms have provided the computational foundations for the rapid growth in applications of statistical machine learning in recent years. There is, however, limited theoretical understanding of the relationships between these 2 kinds of methodology, and limited understanding of relative strengths and weaknesses. Moreover, existing results have been obtained primarily in the setting of convex functions (for optimization) and log-concave functions (for sampling). In this setting, where local properties determine global properties, optimization algorithms are unsurprisingly more efficient computationally than sampling algorithms. We instead examine a class of nonconvex objective functions that arise in mixture modeling and multistable systems. In this nonconvex setting, we find that the computational complexity of sampling algorithms scales linearly with the model dimension while that of optimization algorithms scales exponentially.
KW - Computational complexity
KW - Langevin Monte Carlo
KW - Nonconvex optimization
UR - http://www.scopus.com/inward/record.url?scp=85073310401&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85073310401&partnerID=8YFLogxK
U2 - 10.1073/pnas.1820003116
DO - 10.1073/pnas.1820003116
M3 - Article
C2 - 31570618
AN - SCOPUS:85073310401
SN - 0027-8424
VL - 116
SP - 20881
EP - 20885
JO - Proceedings of the National Academy of Sciences of the United States of America
JF - Proceedings of the National Academy of Sciences of the United States of America
IS - 42
ER -