Support union recovery in high-dimensional multivariate regression
From MaRDI portal
Publication:2429923
DOI10.1214/09-AOS776zbMath1373.62372arXiv0808.0711MaRDI QIDQ2429923
Martin J. Wainwright, Michael I. Jordan, Guillaume Obozinski
Publication date: 5 April 2011
Published in: The Annals of Statistics (Search for Journal in Brave)
Full work available at URL: https://arxiv.org/abs/0808.0711
variable selection; sparsity; group Lasso; multivariate regression; Lasso; second-order cone program; block-norm; high-dimensional scaling; simultaneous Lasso
62J07: Ridge regression; shrinkage estimators (Lasso)
62F07: Statistical ranking and selection procedures
Related Items
CLEAR: Covariant LEAst-Square Refitting with Applications to Image Restoration, Unsupervised learning of compositional sparse code for natural image representation, Structured sparsity through convex optimization, A selective review of group selection in high-dimensional models, A unified framework for high-dimensional analysis of \(M\)-estimators with decomposable regularizers, The benefit of group sparsity in group inference with de-biased scaled group Lasso, A partially linear framework for massive heterogeneous data, The geometry of least squares in the 21st century, ``Grouping strategies and thresholding for high dimensional linear models: discussion, Lasso penalized model selection criteria for high-dimensional multivariate linear regression analysis, Sparse regression and support recovery with \(\mathbb{L}_2\)-boosting algorithms, Robust inference on average treatment effects with possibly more covariates than observations, Estimation of (near) low-rank matrices with noise and high-dimensional scaling, Oracle inequalities and optimal inference under group sparsity, Tree-guided group lasso for multi-response regression with structured sparsity, with an application to eQTL mapping, Tight conditions for consistency of variable selection in the context of high dimensionality, Accuracy guaranties for \(\ell_{1}\) recovery of block-sparse signals, Optimal variable selection in multi-group sparse discriminant analysis, High-dimensional Ising model selection using \(\ell _{1}\)-regularized logistic regression, Regularized multivariate regression for identifying master predictors with application to integrative genomics study of breast cancer, The benefit of group sparsity, Structured variable selection via prior-induced hierarchical penalty functions, On the oracle property of a generalized adaptive elastic-net for multivariate linear regression with a diverging number of parameters, On model selection consistency of regularized M-estimators, Structured regularization for conditional Gaussian graphical models, Prototype selection for parameter estimation in complex models, Structure estimation for discrete graphical models: generalized covariance matrices and their inverses, Functional linear regression for functional response via sparse basis selection, Sparse reduced-rank regression with covariance estimation, Multivariate sparse group lasso for the multivariate multiple linear regression with an arbitrary group structure
Uses Software
Cites Work
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Concentration inequalities and model selection. Ecole d'Eté de Probabilités de Saint-Flour XXXIII -- 2003.
- The benefit of group sparsity
- Lasso-type recovery of sparse representations for high-dimensional data
- The composite absolute penalties family for grouped and hierarchical variable selection
- Adaptive estimation of a quadratic functional by model selection.
- Asymptotics for Lasso-type estimators.
- Least angle regression. (With discussion)
- Simultaneous analysis of Lasso and Dantzig selector
- Asymptotic properties of bridge estimators in sparse high-dimensional regression models
- Variable selection for multicategory SVM via adaptive sup-norm regularization
- High-dimensional graphs and variable selection with the Lasso
- Just relax: convex programming methods for identifying sparse signals in noise
- The Group Lasso for Logistic Regression
- Atomic Decomposition by Basis Pursuit
- Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties
- A new approach to variable selection in least squares problems
- Uncertainty principles and ideal atomic decomposition
- A Statistical View of Some Chemometrics Regression Tools
- Information-Theoretic Limits on Sparsity Recovery in the High-Dimensional and Noisy Setting
- Sharp Thresholds for High-Dimensional and Noisy Sparsity Recovery Using $\ell _{1}$-Constrained Quadratic Programming (Lasso)
- Model Selection and Estimation in Regression with Grouped Variables