Efficient regularized regression with \(L_0\) penalty for variable selection and network construction (Q2011726): Difference between revisions

From MaRDI portal
Importer (talk | contribs)
Created a new Item
 
ReferenceBot (talk | contribs)
Changed an Item
 
(5 intermediate revisions by 5 users not shown)
Property / describes a project that uses
 
Property / describes a project that uses: sparsenet / rank
 
Normal rank
Property / MaRDI profile type
 
Property / MaRDI profile type: MaRDI publication profile / rank
 
Normal rank
Property / full work available at URL
 
Property / full work available at URL: https://doi.org/10.1155/2016/3456153 / rank
 
Normal rank
Property / OpenAlex ID
 
Property / OpenAlex ID: W2532146222 / rank
 
Normal rank
Property / Wikidata QID
 
Property / Wikidata QID: Q37397132 / rank
 
Normal rank
Property / cites work
 
Property / cites work: A new look at the statistical model identification / rank
 
Normal rank
Property / cites work
 
Property / cites work: Estimating the dimension of a model / rank
 
Normal rank
Property / cites work
 
Property / cites work: The risk inflation criterion for multiple regression / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q4864293 / rank
 
Normal rank
Property / cites work
 
Property / cites work: The Adaptive Lasso and Its Oracle Properties / rank
 
Normal rank
Property / cites work
 
Property / cites work: Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties / rank
 
Normal rank
Property / cites work
 
Property / cites work: <i>SparseNet</i>: Coordinate Descent With Nonconvex Penalties / rank
 
Normal rank
Property / cites work
 
Property / cites work: Nearly unbiased variable selection under minimax concave penalty / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q4921683 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Sparse Approximation via Penalty Decomposition Methods / rank
 
Normal rank
Property / cites work
 
Property / cites work: On the adaptive elastic net with a diverging number of parameters / rank
 
Normal rank
Property / cites work
 
Property / cites work: Tuning Parameter Selection in High Dimensional Penalized Likelihood / rank
 
Normal rank
Property / cites work
 
Property / cites work: Improved Iteratively Reweighted Least Squares for Unconstrained Smoothed $\ell_q$ Minimization / rank
 
Normal rank
Property / cites work
 
Property / cites work: Iteratively reweighted least squares minimization for sparse recovery / rank
 
Normal rank
Property / cites work
 
Property / cites work: Partial Correlation Estimation by Joint Sparse Regression Models / rank
 
Normal rank
Property / cites work
 
Property / cites work: Stability Selection / rank
 
Normal rank
Property / cites work
 
Property / cites work: A Generic Path Algorithm for Regularized Statistical Estimation / rank
 
Normal rank
links / mardi / namelinks / mardi / name
 

Latest revision as of 05:09, 14 July 2024

scientific article
Language Label Description Also known as
English
Efficient regularized regression with \(L_0\) penalty for variable selection and network construction
scientific article

    Statements

    Efficient regularized regression with \(L_0\) penalty for variable selection and network construction (English)
    0 references
    0 references
    0 references
    4 August 2017
    0 references
    Summary: Variable selections for regression with high-dimensional big data have found many applications in bioinformatics and computational biology. One appealing approach is the \(L_0\) regularized regression which penalizes the number of nonzero features in the model directly. However, it is well known that \(L_0\) optimization is NP-hard and computationally challenging. In this paper, we propose efficient EM (\(L_0\)EM) and dual \(L_0\)EM (D\(L_0\)EM) algorithms that directly approximate the \(L_0\) optimization problem. While \(L_0\)EM is efficient with large sample size, D\(L_0\)EM is efficient with high-dimensional (\(n \ll m\)) data. They also provide a natural solution to all \(L_p\)\ \ \(p \in [0,2]\) problems, including lasso with \(p = 1\) and elastic net with \(p \in [1,2]\). The regularized parameter \(\lambda\) can be determined through cross validation or AIC and BIC. We demonstrate our methods through simulation and high-dimensional genomic data. The results indicate that \(L_0\) has better performance than lasso, SCAD, and MC+, and \(L_0\) with AIC or BIC has similar performance as computationally intensive cross validation. The proposed algorithms are efficient in identifying the nonzero variables with less bias and constructing biologically important networks with high-dimensional big data.
    0 references
    \(L_0\) regularized regression
    0 references
    high-dimensional big data
    0 references
    EM algorithms
    0 references

    Identifiers

    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references