Finite sample approximation results for principal component analysis: A matrix perturbation approach
Open Access
- 1 December 2008
- journal article
- Published by Institute of Mathematical Statistics in The Annals of Statistics
- Vol. 36 (6) , 2791-2817
- https://doi.org/10.1214/08-aos618
Abstract
Principal component analysis (PCA) is a standard tool for dimensional reduction of a set of n observations (samples), each with p variables. In this paper, using a matrix perturbation approach, we study the nonasymptotic relation between the eigenvalues and eigenvectors of PCA computed on a finite sample of size n, and those of the limiting population PCA as n→∞. As in machine learning, we present a finite sample theorem which holds with high probability for the closeness between the leading eigenvalue and eigenvector of sample PCA and population PCA under a spiked covariance model. In addition, we also consider the relation between finite sample PCA and the asymptotic results in the joint limit p, n→∞, with p/n=c. We present a matrix perturbation view of the “phase transition phenomenon,” and a simple linear-algebra based derivation of the eigenvalue and eigenvector overlap in this asymptotic limit. Moreover, our analysis also applies for finite p, n where we show that although there is no sharp phase transition as in the infinite case, either as a function of noise level or as a function of sample size n, the eigenvector of sample PCA may exhibit a sharp “loss of tracking,” suddenly losing its relation to the (true) eigenvector of the population PCA matrix. This occurs due to a crossover between the eigenvalue due to the signal and the largest eigenvalue due to noise, whose eigenvector points in a random direction.Keywords
All Related Versions
This publication has 22 references indexed in Scilit:
- Tracy–Widom limit for the largest eigenvalue of a large class of complex sample covariance matricesThe Annals of Probability, 2007
- Phase transition of the largest eigenvalue for nonnull complex sample covariance matricesThe Annals of Probability, 2005
- Partial least squares, Beer's law and the net analyte signal: statistical modeling and analysisJournal of Chemometrics, 2005
- On the distribution of the largest eigenvalue in principal components analysisThe Annals of Statistics, 2001
- A Gaussian scenario for unsupervised learningJournal of Physics A: General Physics, 1996
- Optimal unsupervised learningJournal of Physics A: General Physics, 1994
- Statistical mechanics of unsupervised structure recognitionJournal of Physics A: General Physics, 1994
- On Wielandt's Inequality and Its Application to the Asymptotic Distribution of the Eigenvalues of a Random Symmetric MatrixThe Annals of Statistics, 1991
- Asymptotic Theory for Principal Component AnalysisThe Annals of Mathematical Statistics, 1963
- On the Sampling Theory of Roots of Determinantal EquationsThe Annals of Mathematical Statistics, 1939