Support union recovery in high-dimensional multivariate regression
Top Cited Papers
Open Access
- 1 February 2011
- journal article
- Published by Institute of Mathematical Statistics in The Annals of Statistics
- Vol. 39 (1)
- https://doi.org/10.1214/09-aos776
Abstract
In multivariate regression, a $K$-dimensional response vector is regressed upon a common set of $p$ covariates, with a matrix $B^*\in\mathbb{R}^{p\times K}$ of regression coefficients. We study the behavior of the multivariate group Lasso, in which block regularization based on the $\ell_1/\ell_2$ norm is used for support union recovery, or recovery of the set of $s$ rows for which $B^*$ is nonzero. Under high-dimensional scaling, we show that the multivariate group Lasso exhibits a threshold for the recovery of the exact row pattern with high probability over the random design and noise that is specified by the sample complexity parameter $\theta(n,p,s):=n/[2\psi(B^*)\log(p-s)]$. Here $n$ is the sample size, and $\psi(B^*)$ is a sparsity-overlap function measuring a combination of the sparsities and overlaps of the $K$-regression coefficient vectors that constitute the model. We prove that the multivariate group Lasso succeeds for problem sequences $(n,p,s)$ such that $\theta(n,p,s)$ exceeds a critical level $\theta_u$, and fails for sequences such that $\theta(n,p,s)$ lies below a critical level $\theta_{\ell}$. For the special case of the standard Gaussian ensemble, we show that $\theta_{\ell}=\theta_u$ so that the characterization is sharp. The sparsity-overlap function $\psi(B^*)$ reveals that, if the design is uncorrelated on the active rows, $\ell_1/\ell_2$ regularization for multivariate regression never harms performance relative to an ordinary Lasso approach and can yield substantial improvements in sample complexity (up to a factor of $K$) when the coefficient vectors are suitably orthogonal. For more general designs, it is possible for the ordinary Lasso to outperform the multivariate group Lasso. We complement our analysis with simulations that demonstrate the sharpness of our theoretical results, even for relatively small problems.Comment: Published in at http://dx.doi.org/10.1214/09-AOS776 the Annals of Statistics (http://www.imstat.org/aos/) by the Institute of Mathematical Statistics (http://www.imstat.org
Keywords
All Related Versions
This publication has 24 references indexed in Scilit:
- The benefit of group sparsityThe Annals of Statistics, 2010
- The composite absolute penalties family for grouped and hierarchical variable selectionThe Annals of Statistics, 2009
- Sparse Additive ModelsJournal of the Royal Statistical Society Series B: Statistical Methodology, 2009
- Simultaneous analysis of Lasso and Dantzig selectorThe Annals of Statistics, 2009
- Lasso-type recovery of sparse representations for high-dimensional dataThe Annals of Statistics, 2009
- Asymptotic properties of bridge estimators in sparse high-dimensional regression modelsThe Annals of Statistics, 2008
- Simultaneous Variable SelectionTechnometrics, 2005
- Uncertainty principles and ideal atomic decompositionIEEE Transactions on Information Theory, 2001
- Adaptive estimation of a quadratic functional by model selectionThe Annals of Statistics, 2000
- A new approach to variable selection in least squares problemsIMA Journal of Numerical Analysis, 2000