Abstract
We investigate ways in which an algorithm can improve its expected performance by fine-tuning itself automatically with respect to an arbitrary, unknown input distribution. We give such self-improving algorithms for sorting and clustering. The highlights of this work: (i) a sorting algorithm with optimal expected limiting running time; and (ii) a k-median algorithm over the Hamming cube with linear expected limiting running time. In all cases, the algorithm begins with a learning phase during which it adjusts itself to the input distribution (typically in a logarithmic number of rounds), followed by a stationary regime in which the algorithm settles to its optimized incarnation.
| Original language | English (US) |
|---|---|
| Pages | 261-270 |
| Number of pages | 10 |
| DOIs | |
| State | Published - 2006 |
| Externally published | Yes |
| Event | Seventeenth Annual ACM-SIAM Symposium on Discrete Algorithms - Miami, FL, United States Duration: Jan 22 2006 → Jan 24 2006 |
Other
| Other | Seventeenth Annual ACM-SIAM Symposium on Discrete Algorithms |
|---|---|
| Country/Territory | United States |
| City | Miami, FL |
| Period | 1/22/06 → 1/24/06 |
All Science Journal Classification (ASJC) codes
- Software
- General Mathematics
Fingerprint
Dive into the research topics of 'Self-improving algorithms'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver