The smooth-Lasso and other \(\ell _{1}+\ell _{2}\)-penalized methods
From MaRDI portal
Publication:1952223
DOI10.1214/11-EJS638zbMath1274.62443arXiv1003.4885MaRDI QIDQ1952223
Sara van de Geer, Mohamed Hebiri
Publication date: 28 May 2013
Published in: Electronic Journal of Statistics (Search for Journal in Brave)
Full work available at URL: https://arxiv.org/abs/1003.4885
high-dimensional data; variable selection; sparsity; Lasso; elastic-net; restricted eigenvalues; LARS algorithm
62F12: Asymptotic properties of parametric estimators
62J07: Ridge regression; shrinkage estimators (Lasso)
62J05: Linear regression; mixed models
62H20: Measures of association (correlation, canonical correlation, etc.)
Related Items
Unnamed Item, Topology of Pareto Sets of Strongly Convex Problems, Graph-Based Regularization for Regression Problems with Alignment and Highly Correlated Designs, Structured sparse support vector machine with ordered features, A unified class of penalties with the capability of producing a differentiable alternative to l1 norm penalty, On estimation error bounds of the Elastic Net when p ≫ n, Oracle Inequalities for Convex Loss Functions with Nonlinear Targets, The sparse Laplacian shrinkage estimator for high-dimensional regression, Properties and iterative methods for the lasso and its variants, Convex relaxations of penalties for sparse correlated variables with bounded total variation, A coordinate descent algorithm for computing penalized smooth quantile regression, Oracle inequalities for high-dimensional prediction, Adaptive function-on-scalar regression with a smoothing elastic net, Free disposal hull condition to verify when efficiency coincides with weak efficiency, Smoothly adaptively centered ridge estimator, Sparse Laplacian shrinkage with the graphical Lasso estimator for regression problems, Nonparametric estimation of the random coefficients model: an elastic net approach, High-dimensional variable screening and bias in subsequent inference, with an empirical comparison, Perspective maximum likelihood-type estimation via proximal decomposition, Innovated interaction screening for high-dimensional nonlinear classification, An extended variable inclusion and shrinkage algorithm for correlated variables, Structured regularization for conditional Gaussian graphical models, Submodular functions: from discrete to continuous domains
Cites Work
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- The Adaptive Lasso and Its Oracle Properties
- Exponential screening and optimal rates of sparse estimation
- Some theoretical results on the grouped variables Lasso
- Properties and refinements of the fused Lasso
- The sparsity and bias of the LASSO selection in high-dimensional linear regression
- Shrinkage and model selection with correlated variables via weighted fusion
- Lasso-type recovery of sparse representations for high-dimensional data
- Relaxed Lasso
- Least angle regression. (With discussion)
- Honest variable selection in linear and logistic regression models via \(\ell _{1}\) and \(\ell _{1}+\ell _{2}\) penalization
- On the conditions used to prove oracle results for the Lasso
- Simultaneous analysis of Lasso and Dantzig selector
- On the adaptive elastic net with a diverging number of parameters
- Sparsity oracle inequalities for the Lasso
- Sup-norm convergence rate and sign concentration property of Lasso and Dantzig estimators
- \(\ell_1\)-penalized quantile regression in high-dimensional sparse models
- Aggregation for Gaussian regression
- High-dimensional graphs and variable selection with the Lasso
- Square root penalty: Adaption to the margin in classification and in edge estimation
- Piecewise linear regularized solution paths
- Nemirovski's Inequalities Revisited
- p-Values for High-Dimensional Regression
- The Group Lasso for Logistic Regression
- $\ell_1$ Trend Filtering
- Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties
- Sparsity and Smoothness Via the Fused Lasso
- Minimax Rates of Estimation for High-Dimensional Linear Regression Over $\ell_q$-Balls
- Regularization and Variable Selection Via the Elastic Net
- On the Non-Negative Garrotte Estimator
- Aggregation by Exponential Weighting and Sharp Oracle Inequalities
- Model Selection and Estimation in Regression with Grouped Variables