TY - JOUR
T1 - An ∞ eigenvector perturbation bound and its application to robust covariance estimation
AU - Fan, Jianqing
AU - Wang, Weichen
AU - Zhong, Yiqiao
N1 - Funding Information:
We would like to acknowledge support for this project from the following grants: NSF grants DMS-1406266, DMS-1662139, DMS-1712591 and NIH grant R01-GM072611-13.
Publisher Copyright:
© 2018 Jianqing Fan, Weichen Wang and Yiqiao Zhong.
PY - 2018/4/1
Y1 - 2018/4/1
N2 - In statistics and machine learning, we are interested in the eigenvectors (or singular vectors) of certain matrices (e.g. covariance matrices, data matrices, etc). However, those matrices are usually perturbed by noises or statistical errors, either from random sampling or structural patterns. The Davis-Kahan sin θ theorem is often used to bound the difference between the eigenvectors of a matrix A and those of a perturbed matrix A = A + E, in terms of 2 norm. In this paper, we prove that when A is a low-rank and incoherent matrix, the ∞ norm perturbation bound of singular vectors (or eigenvectors in the symmetric case) is smaller by a factor of d1 or d2 for left and right vectors, where d1 and d2 are the matrix dimensions. The power of this new perturbation result is shown in robust covariance estimation, particularly when random variables have heavy tails. There, we propose new robust covariance estimators and establish their asymptotic properties using the newly developed perturbation bound. Our theoretical results are verified through extensive numerical experiments.
AB - In statistics and machine learning, we are interested in the eigenvectors (or singular vectors) of certain matrices (e.g. covariance matrices, data matrices, etc). However, those matrices are usually perturbed by noises or statistical errors, either from random sampling or structural patterns. The Davis-Kahan sin θ theorem is often used to bound the difference between the eigenvectors of a matrix A and those of a perturbed matrix A = A + E, in terms of 2 norm. In this paper, we prove that when A is a low-rank and incoherent matrix, the ∞ norm perturbation bound of singular vectors (or eigenvectors in the symmetric case) is smaller by a factor of d1 or d2 for left and right vectors, where d1 and d2 are the matrix dimensions. The power of this new perturbation result is shown in robust covariance estimation, particularly when random variables have heavy tails. There, we propose new robust covariance estimators and establish their asymptotic properties using the newly developed perturbation bound. Our theoretical results are verified through extensive numerical experiments.
KW - Approximate factor model
KW - Incoherence
KW - Low-rank matrices
KW - Matrix perturbation theory
KW - Sparsity
UR - http://www.scopus.com/inward/record.url?scp=85048935983&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85048935983&partnerID=8YFLogxK
M3 - Article
AN - SCOPUS:85048935983
SN - 1532-4435
VL - 18
SP - 1
EP - 42
JO - Journal of Machine Learning Research
JF - Journal of Machine Learning Research
ER -