Penalised robust estimators for sparse and high-dimensional linear models
Publication:2664993
DOI10.1007/S10260-020-00511-ZzbMath1474.62182OpenAlexW3003886944MaRDI QIDQ2664993
Anestis Antoniadis, Umberto Amato, Irène Gijbels, Italia De Feis
Publication date: 18 November 2021
Published in: Statistical Methods and Applications (Search for Journal in Brave)
Full work available at URL: https://lirias.kuleuven.be/handle/123456789/649460
outliersregularizationvariable selectioncontaminationhigh-dimensional regressionwavelet thresholdingnonconvex penalties
Nonparametric regression and quantile regression (62G08) Estimation in multivariate analysis (62H12) Nonparametric robustness (62G35)
Related Items (5)
Uses Software
Cites Work
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Nearly unbiased variable selection under minimax concave penalty
- The Adaptive Lasso and Its Oracle Properties
- Penalized likelihood regression for generalized linear models with non-quadratic penalties
- The \(L_1\) penalized LAD estimator for high dimensional linear regression
- Weighted LAD-LASSO method for robust parameter estimation and variable selection in regression
- Optimal computational and statistical rates of convergence for sparse nonconvex learning problems
- Introduction to robust and quasi-robust statistical methods
- \(\ell_{1}\)-penalization for mixture regression models
- Comments on: \(\ell_{1}\)-penalization for mixture regression models
- High-dimensional regression with noisy and missing data: provable guarantees with nonconvexity
- Wavelet estimation of partially linear models
- Wavelet methods in statistics: some recent developments and their applications
- Lasso-type recovery of sparse representations for high-dimensional data
- Hedonic housing prices and the demand for clean air
- Robust nonnegative garrote variable selection in linear regression
- Robust and sparse estimators for linear regression models
- Asymptotics for Lasso-type estimators.
- Least angle regression. (With discussion)
- Sparse least trimmed squares regression for analyzing high-dimensional large data sets
- Robust regression through the Huber's criterion and adaptive lasso penalty
- Statistical consistency and asymptotic normality for high-dimensional robust \(M\)-estimators
- Robust elastic net estimators for variable selection and identification of proteomic biomarkers
- The power of monitoring: how to make the most of a contaminated multivariate sample
- Simultaneous analysis of Lasso and Dantzig selector
- High-dimensional generalized linear models and the lasso
- Sparsity oracle inequalities for the Lasso
- \(\ell_1\)-penalized quantile regression in high-dimensional sparse models
- New Robust Variable Selection Methods for Linear Regression Models
- Confidence Intervals and Hypothesis Testing for High-Dimensional Regression
- Outlier Detection Using Nonconvex Penalized Regression
- Square-root lasso: pivotal recovery of sparse signals via conic programming
- Scaled sparse linear regression
- Pushing the Limits of Contemporary Statistics: Contributions in Honor of Jayanta K. Ghosh
- Unified LASSO Estimation by Least Squares Approximation
- Robust Linear Model Selection Based on Least Angle Regression
- Asymptotic behavior of general M-estimates for regression and scale with random carriers
- Techniques for nonlinear least squares and robust regression
- Regularization of Wavelet Approximations
- Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties
- Consistency and robustness properties of the S-nonnegative garrote estimator
- Penalized Composite Quasi-Likelihood for Ultrahigh Dimensional Variable Selection
- Fully efficient robust estimation, outlier detection, and variable selection via penalized regression
- Information-Theoretic Limits on Sparsity Recovery in the High-Dimensional and Noisy Setting
- Regularization and Variable Selection Via the Elastic Net
- Penalized partially linear models using sparse representations with an application to fMRI time series
- Estimation of High Dimensional Mean Regression in the Absence of Symmetry and Light Tail Assumptions
- Robust and consistent variable selection in high-dimensional generalized linear models
- Regularized M-estimators with nonconvexity: Statistical and algorithmic theory for local optima
- Robust Estimation of a Location Parameter
- Robust Statistics
- A unified framework for high-dimensional analysis of \(M\)-estimators with decomposable regularizers
- Comments on: \(\ell _{1}\)-penalization for mixture regression models
This page was built for publication: Penalised robust estimators for sparse and high-dimensional linear models