Abstract
A new universal estimator of divergence is presented for multidimensional continuous densities based on κ-nearest-neighbor (κ-NN) distances. Assuming independent and identically distributed (i.i.d.) samples, the new estimator is proved to be asymptotically unbiased and mean-square consistent. In experiments with high-dimensional data, the κ-NN approach generally exhibits faster convergence than previous algorithms. It is also shown that the speed of convergence of the κ-NN method can be further improved by an adaptive choice of κ.
Original language | English (US) |
---|---|
Pages (from-to) | 2392-2405 |
Number of pages | 14 |
Journal | IEEE Transactions on Information Theory |
Volume | 55 |
Issue number | 5 |
DOIs | |
State | Published - 2009 |
All Science Journal Classification (ASJC) codes
- Information Systems
- Computer Science Applications
- Library and Information Sciences
Keywords
- Divergence
- Information measure
- Kullback
- Leibler
- Nearest-neighbor
- Partition
- Random vector
- Universal estimation