A general theory of concave regularization for high-dimensional sparse estimation problems

From MaRDI portal
Publication:5965310

DOI10.1214/12-STS399zbMath1331.62353arXiv1108.4988MaRDI QIDQ5965310

Tong Zhang, Cun-Hui Zhang

Publication date: 3 March 2016

Published in: Statistical Science (Search for Journal in Brave)

Full work available at URL: https://arxiv.org/abs/1108.4988



Related Items

Robust Signal Recovery for High-Dimensional Linear Log-Contrast Models with Compositional Covariates, A review of distributed statistical inference, REMI: REGRESSION WITH MARGINAL INFORMATION AND ITS APPLICATION IN GENOME-WIDE ASSOCIATION STUDIES, Regularized projection score estimation of treatment effects in high-dimensional quantile regression, Bayesian Estimation of Gaussian Conditional Random Fields, Fitting sparse linear models under the sufficient and necessary condition for model identification, Global solutions to folded concave penalized nonconvex learning, Best subset selection via a modern optimization lens, \(\ell_0\)-regularized high-dimensional accelerated failure time model, Sparse recovery via nonconvex regularized \(M\)-estimators over \(\ell_q\)-balls, On the strong oracle property of concave penalized estimators with infinite penalty derivative at the origin, GSDAR: a fast Newton algorithm for \(\ell_0\) regularized generalized linear models with statistical guarantee, De-biasing the Lasso with degrees-of-freedom adjustment, Random subspace method for high-dimensional regression with the \texttt{R} package \texttt{regRSM}, High-Dimensional Learning Under Approximate Sparsity with Applications to Nonsmooth Estimation and Regularized Neural Networks, Hard Thresholding Regularised Logistic Regression: Theory and Algorithms, Nonlinear Variable Selection via Deep Neural Networks, Distributed testing and estimation under sparse high dimensional models, Bias versus non-convexity in compressed sensing, The Spike-and-Slab LASSO, Variable selection and parameter estimation with the Atan regularization method, Homogeneity detection for the high-dimensional generalized linear model, Principal components adjusted variable screening, The use of random-effect models for high-dimensional variable selection problems, Conditional sure independence screening by conditional marginal empirical likelihood, Balanced estimation for high-dimensional measurement error models, In defense of LASSO, Oracle inequalities for the lasso in the Cox model, Almost sure uniqueness of a global minimum without convexity, Relaxed sparse eigenvalue conditions for sparse estimation via non-convex regularized regression, A doubly sparse approach for group variable selection, Folded concave penalized sparse linear regression: sparsity, statistical performance, and algorithmic theory for local solutions, Quantile regression for additive coefficient models in high dimensions, On high-dimensional Poisson models with measurement error: hypothesis testing for nonlinear nonconvex optimization, Simultaneous feature selection and outlier detection with optimality guarantees, Sparse signal reconstruction via the approximations of \(\ell_0\) quasinorm, L 0 -regularization for high-dimensional regression with corrupted data, Sparse and robust estimation with ridge minimax concave penalty, Adaptive bridge estimator for Cox model with a diverging number of parameters, Subspace learning by \(\ell^0\)-induced sparsity, A convex-Nonconvex strategy for grouped variable selection, Matrix completion with nonconvex regularization: spectral operators and scalable algorithms, Retire: robust expectile regression in high dimensions, Robust High-Dimensional Regression with Coefficient Thresholding and Its Application to Imaging Data Analysis, Nonconvex penalized reduced rank regression and its oracle properties in high dimensions, Model selection in high-dimensional quantile regression with seamless \(L_0\) penalty, Communication-efficient distributed estimation for high-dimensional large-scale linear regression, Sparse estimation via lower-order penalty optimization methods in high-dimensional linear regression, High-dimensional composite quantile regression: optimal statistical guarantees and fast algorithms, Goodness-of-Fit Tests for High Dimensional Linear Models, Calibrating nonconvex penalized regression in ultra-high dimension, Estimation and inference for precision matrices of nonstationary time series, On the finite-sample analysis of \(\Theta\)-estimators, An unbiased approach to compressed sensing, Estimation and variable selection with exponential weights, Time-varying Hazards Model for Incorporating Irregularly Measured, High-Dimensional Biomarkers, A two-stage regularization method for variable selection and forecasting in high-order interaction model, High-dimensional grouped folded concave penalized estimation via the LLA algorithm, On the finite-sample analysis of \(\Theta\)-estimators, Optimal computational and statistical rates of convergence for sparse nonconvex learning problems, A Tuning-free Robust and Efficient Approach to High-dimensional Regression, Going beyond oracle property: selection consistency and uniqueness of local solution of the generalized linear model, Robust low-rank multiple kernel learning with compound regularization, I-LAMM for sparse learning: simultaneous control of algorithmic complexity and statistical error, Truncated $L^1$ Regularized Linear Regression: Theory and Algorithm, Penalized least squares estimation with weakly dependent data, Tuning parameter selection for the adaptive LASSO in the autoregressive model, Confidence Intervals for Low Dimensional Parameters in High Dimensional Linear Models, Learning latent variable Gaussian graphical model for biomolecular network with low sample complexity, Unnamed Item, On a monotone scheme for nonconvex nonsmooth optimization with applications to fracture mechanics, Tractable ADMM schemes for computing KKT points and local minimizers for \(\ell_0\)-minimization problems, Sorted concave penalized regression, Strong oracle optimality of folded concave penalized estimation, Endogeneity in high dimensions, Bayesian Bootstrap Spike-and-Slab LASSO, A unified primal dual active set algorithm for nonconvex sparse recovery, Introduction to the special issue on sparsity and regularization methods, Estimating structured high-dimensional covariance and precision matrices: optimal rates and adaptive estimation, Variance prior forms for high-dimensional Bayesian variable selection, OR Forum—An Algorithmic Approach to Linear Regression, Iteratively reweighted \(\ell_1\)-penalized robust regression, Asymptotic normality and optimalities in estimation of large Gaussian graphical models, High-dimensional linear model selection motivated by multiple testing, Majorized proximal alternating imputation for regularized rank constrained matrix completion, Second-order Stein: SURE for SURE and other applications in high-dimensional inference, Dynamic variable selection with spike-and-slab process priors, A Simple Method for Estimating Interactions Between a Treatment and a Large Number of Covariates, Simultaneous Variable and Covariance Selection With the Multivariate Spike-and-Slab LASSO, Nonnegative estimation and variable selection under minimax concave penalty for sparse high-dimensional linear regression models, Smoothing Newton method for \(\ell^0\)-\(\ell^2\) regularized linear inverse problem, A theoretical understanding of self-paced learning, Fast Best Subset Selection: Coordinate Descent and Local Combinatorial Optimization Algorithms, Accelerated Stochastic Algorithms for Nonconvex Finite-Sum and Multiblock Optimization, Weighted thresholding homotopy method for sparsity constrained optimization, Nonbifurcating Phylogenetic Tree Inference via the Adaptive LASSO, Joint feature screening for ultra-high-dimensional sparse additive hazards model by the sparsity-restricted pseudo-score estimator, Sample average approximation with sparsity-inducing penalty for high-dimensional stochastic programming, Robust moderately clipped LASSO for simultaneous outlier detection and variable selection, High-dimensional linear regression with hard thresholding regularization: theory and algorithm


Uses Software


Cites Work