Least angle and \(\ell _{1}\) penalized regression: a review
From MaRDI portal
Publication:975564
DOI10.1214/08-SS035zbMath1189.62070arXiv0802.0964MaRDI QIDQ975564
Chris Fraley, Lukas Meier, Nam Hee Choi, T. C. Hesterberg
Publication date: 9 June 2010
Published in: Statistics Surveys (Search for Journal in Brave)
Full work available at URL: https://arxiv.org/abs/0802.0964
62G08: Nonparametric regression and quantile regression
62J07: Ridge regression; shrinkage estimators (Lasso)
62J99: Linear inference, regression
Related Items
A variable selection proposal for multiple linear regression analysis, Absolute penalty and shrinkage estimation in partially linear models, Weighted LAD-LASSO method for robust parameter estimation and variable selection in regression, Sparsity with sign-coherent groups of variables via the cooperative-Lasso, Graph optimization for dimensionality reduction with sparsity constraints, Variable selection using penalized empirical likelihood, A survey of cross-validation procedures for model selection, Probabilities of discrepancy between minima of cross-validation, Vapnik bounds and true risks, Estimation of a semiparametric recursive bivariate probit model in the presence of endogeneity
Uses Software
Cites Work
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Sparse inverse covariance estimation with the graphical lasso
- The Adaptive Lasso and Its Oracle Properties
- LASSO-pattern search algorithm with application to ophthalmology and genomic data
- Component selection and smoothing in multivariate nonparametric regression
- Discussion: One-step sparse estimates in nonconcave penalized likelihood models
- Rejoinder: One-step sparse estimates in nonconcave penalized likelihood models
- The sparsity and bias of the LASSO selection in high-dimensional linear regression
- Lasso-type recovery of sparse representations for high-dimensional data
- Parsimonious additive models
- DALASS: variable selection in discriminant analysis via the LASSO
- Relaxed Lasso
- Input selection and shrinkage in multiresponse linear regression
- Multivariate adaptive regression splines
- A decision-theoretic generalization of on-line learning and an application to boosting
- Persistene in high-dimensional linear predictor-selection and the virtue of overparametrization
- Least angle regression. (With discussion)
- High-dimensional generalized linear models and the lasso
- Forward stagewise regression and the monotone lasso
- Sparsity oracle inequalities for the Lasso
- Smoothing \(\ell_1\)-penalized estimators for high-dimensional time-course data
- Pathwise coordinate optimization
- On the ``degrees of freedom of the lasso
- The Dantzig selector: statistical estimation when \(p\) is much larger than \(n\). (With discussions and rejoinder).
- High-dimensional graphs and variable selection with the Lasso
- Piecewise linear regularized solution paths
- Better Subset Regression Using the Nonnegative Garrote
- Model selection and estimation in the Gaussian graphical model
- The Group Lasso for Logistic Regression
- Unified LASSO Estimation by Least Squares Approximation
- Robust Linear Model Selection Based on Least Angle Regression
- Regressions by Leaps and Bounds
- Atomic Decomposition by Basis Pursuit
- Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties
- A new approach to variable selection in least squares problems
- Sparsity and Smoothness Via the Fused Lasso
- A Statistical View of Some Chemometrics Regression Tools
- Regularization and Variable Selection Via the Elastic Net
- On the Non-Negative Garrotte Estimator
- Model Selection and Estimation in Regression with Grouped Variables
- Adaptive Lasso for Cox's proportional hazards model
- Covariance selection for nonchordal graphs via chordal embedding
- Variable Selection and Model Building via Likelihood Basis Pursuit
- Numerical Methods of Statistics
- The elements of statistical learning. Data mining, inference, and prediction
- Structural modelling with sparse kernels