Sure Independence Screening for Ultrahigh Dimensional Feature Space
Top Cited Papers
Open Access
- 3 October 2008
- journal article
- Published by Oxford University Press (OUP) in Journal of the Royal Statistical Society Series B: Statistical Methodology
- Vol. 70 (5) , 849-911
- https://doi.org/10.1111/j.1467-9868.2008.00674.x
Abstract
Summary. Variable selection plays an important role in high dimensional statistical modelling which nowadays appears in many areas and is key to various scientific discoveries. For problems of large scale or dimensionality p, accuracy of estimation and computational cost are two top concerns. Recently, Candes and Tao have proposed the Dantzig selector using L1-regularization and showed that it achieves the ideal risk up to a logarithmic factor log (p). Their innovative procedure and remarkable result are challenged when the dimensionality is ultrahigh as the factor log (p) can be large and their uniform uncertainty principle can fail. Motivated by these concerns, we introduce the concept of sure screening and propose a sure screening method that is based on correlation learning, called sure independence screening, to reduce dimensionality from high to a moderate scale that is below the sample size. In a fairly general asymptotic framework, correlation learning is shown to have the sure screening property for even exponentially growing dimensionality. As a methodological extension, iterative sure independence screening is also proposed to enhance its finite sample performance. With dimension reduced accurately from high to below sample size, variable selection can be improved on both speed and accuracy, and can then be accomplished by a well-developed method such as smoothly clipped absolute deviation, the Dantzig selector, lasso or adaptive lasso. The connections between these penalized least squares methods are also elucidated.Keywords
All Related Versions
Funding Information
- National Science Foundation (DMS-0354223, DMS-0704337, DMS-0714554, DMS-0806030)
- National Institutes of Health (R01-GM072611)
This publication has 86 references indexed in Scilit:
- Multivariate analysis and Jacobi ensembles: Largest eigenvalue, Tracy–Widom limits and rates of convergenceThe Annals of Statistics, 2008
- High-dimensional classification using features annealed independence rulesThe Annals of Statistics, 2008
- Tuning parameter selectors for the smoothly clipped absolute deviation methodBiometrika, 2007
- Fisher Lecture: Dimension Reduction in RegressionStatistical Science, 2007
- Boosting for high-dimensional linear modelsThe Annals of Statistics, 2006
- Statistical significance for genomewide studiesProceedings of the National Academy of Sciences, 2003
- Better Subset Regression Using the Nonnegative GarroteTechnometrics, 1995
- Minimax risk overl p -balls forl p -errorProbability Theory and Related Fields, 1994
- A Statistical View of Some Chemometrics Regression ToolsTechnometrics, 1993
- Sliced Inverse Regression for Dimension ReductionJournal of the American Statistical Association, 1991