• DocumentCode
    750087
  • Title

    Sharp Thresholds for High-Dimensional and Noisy Sparsity Recovery Using \\ell _{1} -Constrained Quadratic Programming (Lasso)

  • Author

    Wainwright, Martin J.

  • Author_Institution
    Dept. of Stat., Univ. of California, Berkeley, CA
  • Volume
    55
  • Issue
    5
  • fYear
    2009
  • fDate
    5/1/2009 12:00:00 AM
  • Firstpage
    2183
  • Lastpage
    2202
  • Abstract
    The problem of consistently estimating the sparsity pattern of a vector beta* isin Rp based on observations contaminated by noise arises in various contexts, including signal denoising, sparse approximation, compressed sensing, and model selection. We analyze the behavior of l1-constrained quadratic programming (QP), also referred to as the Lasso, for recovering the sparsity pattern. Our main result is to establish precise conditions on the problem dimension p, the number k of nonzero elements in beta*, and the number of observations n that are necessary and sufficient for sparsity pattern recovery using the Lasso. We first analyze the case of observations made using deterministic design matrices and sub-Gaussian additive noise, and provide sufficient conditions for support recovery and linfin-error bounds, as well as results showing the necessity of incoherence and bounds on the minimum value. We then turn to the case of random designs, in which each row of the design is drawn from a N (0, Sigma) ensemble. For a broad class of Gaussian ensembles satisfying mutual incoherence conditions, we compute explicit values of thresholds 0 < thetasl(Sigma) les thetasu(Sigma) < +infin with the following properties: for any delta > 0, if n > 2 (thetasu + delta) klog (p- k), then the Lasso succeeds in recovering the sparsity pattern with probability converging to one for large problems, whereas for n < 2 (thetasl - delta)klog (p - k), then the probability of successful recovery converges to zero. For the special case of the uniform Gaussian ensemble (Sigma = Iptimesp), we show that thetasl = thetas<u = 1, so that the precise threshold n = 2 klog(p- k) is exactly determined.
  • Keywords
    quadratic programming; signal denoising; sparse matrices; Lasso; compressed sensing; deterministic design matrices; high-dimensional recovery; l1 -constrained quadratic programming; model selection; noisy sparsity recovery; sharp thresholds; signal denoising; sparse approximation; sparsity pattern recovery; subGaussian additive noise; uniform Gaussian ensemble; Additive noise; Compressed sensing; Context modeling; Graphical models; Pattern analysis; Polynomials; Quadratic programming; Signal denoising; Statistics; Sufficient conditions; $ell _{1}$-constraints; Compressed sensing; convex relaxation; high-dimensional inference; model selection; phase transitions; signal denoising; sparse approximation; subset selection;
  • fLanguage
    English
  • Journal_Title
    Information Theory, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    0018-9448
  • Type

    jour

  • DOI
    10.1109/TIT.2009.2016018
  • Filename
    4839045