The adaptive and the thresholded Lasso for potentially misspecified models (and a lower bound for the Lasso)

From MaRDI portal
Publication:1952206


DOI10.1214/11-EJS624zbMath1274.62471arXiv1001.5176MaRDI QIDQ1952206

Shuheng Zhou, Sara van de Geer, Peter Bühlmann

Publication date: 28 May 2013

Published in: Electronic Journal of Statistics (Search for Journal in Brave)

Full work available at URL: https://arxiv.org/abs/1001.5176


62G08: Nonparametric regression and quantile regression

62J07: Ridge regression; shrinkage estimators (Lasso)


Related Items

Unnamed Item, Variable Selection With Second-Generation P-Values, Model Selection With Lasso-Zero: Adding Straw to the Haystack to Better Find Needles, Robust recovery of signals with partially known support information using weighted BPDN, Quasi-likelihood and/or robust estimation in high dimensions, Discussion of: ``Grouping strategies and thresholding for high dimension linear models, A Critical Review of LASSO and Its Derivatives for Variable Selection Under Dependence Among Covariates, A Unified Framework for Change Point Detection in High-Dimensional Linear Models, Weak Signal Identification and Inference in Penalized Likelihood Models for Categorical Responses, Testing stochastic dominance with many conditioning variables, A power analysis for Model-X knockoffs with \(\ell_p\)-regularized statistics, Positive-definite thresholding estimators of covariance matrices with zeros, An integrated surrogate model constructing method: annealing combinable Gaussian process, Regularized estimation in sparse high-dimensional time series models, Orthogonal one step greedy procedure for heteroscedastic linear models, Best subset selection via a modern optimization lens, Thresholding least-squares inference in high-dimensional regression models, \(\ell_{0}\)-penalized maximum likelihood for sparse directed acyclic graphs, Statistical significance in high-dimensional linear models, CAM: causal additive models, high-dimensional order search and penalized regression, Regularized rank-based estimation of high-dimensional nonparanormal graphical models, Endogenous treatment effect estimation using high-dimensional instruments and double selection, Efficient estimation of approximate factor models via penalized maximum likelihood, Monotone splines Lasso, High-dimensional simultaneous inference with the bootstrap, High-dimensional inference: confidence intervals, \(p\)-values and R-software \texttt{hdi}, High-dimensional variable selection via low-dimensional adaptive learning, Ridge regression revisited: debiasing, thresholding and bootstrap, Rejoinder on: ``Hierarchical inference for genome-wide association studies: a view on methodology with software, A look at robustness and stability of \(\ell_1\)-versus \(\ell_0\)-regularization: discussion of papers by Bertsimas et al. and Hastie et al., High-dimensional variable screening and bias in subsequent inference, with an empirical comparison, Lasso and probabilistic inequalities for multivariate point processes, Preconditioning the Lasso for sign consistency, D-trace estimation of a precision matrix using adaptive lasso penalties, Calibrating nonconvex penalized regression in ultra-high dimension, An1-oracle inequality for the Lasso in multivariate finite mixture of multivariate Gaussian regression models, Estimation for High-Dimensional Linear Mixed-Effects Models Using ℓ1-Penalization


Uses Software


Cites Work