Learning to Warm-Start Fixed-Point Optimization Algorithms

Research output: Contribution to journalArticlepeer-review

13 Scopus citations

Abstract

We introduce a machine-learning framework to warm-start fixed-point optimization algorithms. Our architecture consists of a neural network mapping problem parameters to warm starts, followed by a predefined number of fixed-point iterations. We propose two loss functions designed to either minimize the fixed-point residual or the distance to a ground truth solution. In this way, the neural network predicts warm starts with the end-to-end goal of minimizing the downstream loss. An important feature of our architecture is its flexibility, in that it can predict a warm start for fixed-point algorithms run for any number of steps, without being limited to the number of steps it has been trained on. We provide PAC-Bayes generalization bounds on unseen data for common classes of fixed-point operators: contractive, linearly convergent, and averaged. Applying this framework to well-known applications in control, statistics, and signal processing, we observe a significant reduction in the number of iterations and solution time required to solve these problems, through learned warm starts.

Original languageEnglish (US)
JournalJournal of Machine Learning Research
Volume25
StatePublished - 2024
Externally publishedYes

All Science Journal Classification (ASJC) codes

  • Software
  • Control and Systems Engineering
  • Statistics and Probability
  • Artificial Intelligence

Keywords

  • fixed-point problems
  • generalization bounds
  • learning to optimize
  • parametric convex optimization
  • warm start

Fingerprint

Dive into the research topics of 'Learning to Warm-Start Fixed-Point Optimization Algorithms'. Together they form a unique fingerprint.

Cite this