TY - JOUR
T1 - Verification of first-order methods for parametric quadratic optimization
AU - Ranjan, Vinit
AU - Stellato, Bartolomeo
N1 - Publisher Copyright:
© Springer-Verlag GmbH Germany, part of Springer Nature and Mathematical Optimization Society 2025.
PY - 2025
Y1 - 2025
N2 - We introduce a numerical framework to verify the finite step convergence of first-order methods for parametric convex quadratic optimization. We formulate the verification problem as a mathematical optimization problem where we maximize a performance metric (e.g., fixed-point residual at the last iteration) subject to constraints representing proximal algorithm steps (e.g., linear system solutions, projections, or gradient steps). Our framework is highly modular because we encode a wide range of proximal algorithms as variations of two primitive steps: affine steps and element-wise maximum steps. Compared to standard convergence analysis and performance estimation techniques, we can explicitly quantify the effects of warm-starting by directly representing the sets where the initial iterates and parameters live. We show that the verification problem is NP-hard, and we construct strong semidefinite programming relaxations using various constraint tightening techniques. Numerical examples in nonnegative least squares, network utility maximization, Lasso, and optimal control show a significant reduction in pessimism of our framework compared to standard worst-case convergence analysis techniques.
AB - We introduce a numerical framework to verify the finite step convergence of first-order methods for parametric convex quadratic optimization. We formulate the verification problem as a mathematical optimization problem where we maximize a performance metric (e.g., fixed-point residual at the last iteration) subject to constraints representing proximal algorithm steps (e.g., linear system solutions, projections, or gradient steps). Our framework is highly modular because we encode a wide range of proximal algorithms as variations of two primitive steps: affine steps and element-wise maximum steps. Compared to standard convergence analysis and performance estimation techniques, we can explicitly quantify the effects of warm-starting by directly representing the sets where the initial iterates and parameters live. We show that the verification problem is NP-hard, and we construct strong semidefinite programming relaxations using various constraint tightening techniques. Numerical examples in nonnegative least squares, network utility maximization, Lasso, and optimal control show a significant reduction in pessimism of our framework compared to standard worst-case convergence analysis techniques.
KW - Convergence analysis
KW - First-order methods
KW - Parametric quadratic optimization
KW - Proximal algorithms
KW - Semidefinite relaxations
UR - https://www.scopus.com/pages/publications/105012202970
UR - https://www.scopus.com/inward/citedby.url?scp=105012202970&partnerID=8YFLogxK
U2 - 10.1007/s10107-025-02261-w
DO - 10.1007/s10107-025-02261-w
M3 - Article
AN - SCOPUS:105012202970
SN - 0025-5610
JO - Mathematical Programming
JF - Mathematical Programming
ER -