TY - GEN
T1 - A generalization of principal component analysis to the exponential family
AU - Collins, Michael
AU - Dasgupta, Sanjoy
AU - Schapire, Robert E.
PY - 2002
Y1 - 2002
N2 - Principal component analysis (PCA) is a commonly applied technique for dimensionality reduction. PCA implicitly minimizes a squared loss function, which may be inappropriate for data that is not real-valued, such as binary-valued data. This paper draws on ideas from the Exponential family, Generalized linear models, and Bregman distances, to give a generalization of PCA to loss functions that we argue are better suited to other data types. We describe algorithms for minimizing the loss functions, and give examples on simulated data.
AB - Principal component analysis (PCA) is a commonly applied technique for dimensionality reduction. PCA implicitly minimizes a squared loss function, which may be inappropriate for data that is not real-valued, such as binary-valued data. This paper draws on ideas from the Exponential family, Generalized linear models, and Bregman distances, to give a generalization of PCA to loss functions that we argue are better suited to other data types. We describe algorithms for minimizing the loss functions, and give examples on simulated data.
UR - http://www.scopus.com/inward/record.url?scp=84899014910&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84899014910&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:84899014910
SN - 0262042088
SN - 9780262042086
T3 - Advances in Neural Information Processing Systems
BT - Advances in Neural Information Processing Systems 14 - Proceedings of the 2001 Conference, NIPS 2001
PB - Neural information processing systems foundation
T2 - 15th Annual Neural Information Processing Systems Conference, NIPS 2001
Y2 - 3 December 2001 through 8 December 2001
ER -