Abstract
The performance of on-line algorithms for learning dichotomies is studied. In on-line learning, the number of examples P is equivalent to the learning time, since each example is presented only once. The learning curve, or generalization error as a function of P, depends on the schedule at which the learning rate is lowered. For a target that is a perceptron rule, the learning curve of the perceptron algorithm can decrease as fast as P-1, if the schedule is optimized. If the target is not realizable by a perceptron, the perceptron algorithm does not generally converge to the solution with lowest generalization error. For the case of unrealizability due to a simple output noise, we propose a new on-line algorithm for a perceptron yielding a learning curve that can approach the optimal generalization error as fast as P-1/2. We then generalize the perceptron algorithm to any class of thresholded smooth functions learning a target from that class. For "well-behaved" input distributions, if this algorithm converges to the optimal solution, its learning curve can decrease as fast as P-1.
| Original language | English (US) |
|---|---|
| Pages (from-to) | 303-310 |
| Number of pages | 8 |
| Journal | Advances in Neural Information Processing Systems |
| Volume | 7 |
| State | Published - 1994 |
| Externally published | Yes |
| Event | 7th Advances in Neural Information Processing Systems, NIPS 1994 - Denver, United States Duration: Nov 28 1994 → Dec 1 1994 |
All Science Journal Classification (ASJC) codes
- Signal Processing
- Information Systems
- Computer Networks and Communications
Fingerprint
Dive into the research topics of 'On-line Learning of Dichotomies'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver