Abstract
Linear optimization is many times algorithmically simpler than nonlinear convex optimization. Linear optimization over matroid polytopes, matching polytopes, and path polytopes are examples of problems for which we have simple and efficient combinatorial algorithms but whose nonlinear convex counterpart is harder and admits significantly less efficient algorithms. This motivates the computational model of convex optimization, including the offline, online, and stochastic settings, using a linear optimization oracle. In this computational model we give several new results that improve on the previous state of the art. Our main result is a novel conditional gradient algorithm for smooth and strongly convex optimization over polyhedral sets that performs only a single linear optimization step over the domain on each iteration and enjoys a linear convergence rate. This gives an exponential improvement in convergence rate over previous results. Based on this new conditional gradient algorithm we give the first algorithms for online convex optimization over polyhedral sets that perform only a single linear optimization step over the domain while having optimal regret guarantees, answering an open question of Kalai and Vempala and of Hazan and Kale. Our online algorithms also imply conditional gradient algorithms for nonsmooth and stochastic convex optimization with the same convergence rates as projected (sub)gradient methods.
Original language | English (US) |
---|---|
Pages (from-to) | 1493-1528 |
Number of pages | 36 |
Journal | SIAM Journal on Optimization |
Volume | 26 |
Issue number | 3 |
DOIs | |
State | Published - 2016 |
All Science Journal Classification (ASJC) codes
- Software
- Theoretical Computer Science
- Applied Mathematics
Keywords
- Conditional gradient methods
- Firstorder methods
- Frank-Wolfe algorithm
- Linear programming
- Online convex optimization
- Online learning
- Stochastic optimization