I-LAMM for sparse learning: simultaneous control of algorithmic complexity and statistical error (Q1750288): Difference between revisions

From MaRDI portal
ReferenceBot (talk | contribs)
Changed an Item
Set OpenAlex properties.
 
Property / OpenAlex ID
 
Property / OpenAlex ID: W2964248738 / rank
 
Normal rank

Latest revision as of 10:24, 30 July 2024

scientific article
Language Label Description Also known as
English
I-LAMM for sparse learning: simultaneous control of algorithmic complexity and statistical error
scientific article

    Statements

    I-LAMM for sparse learning: simultaneous control of algorithmic complexity and statistical error (English)
    0 references
    0 references
    0 references
    0 references
    0 references
    18 May 2018
    0 references
    This paper proposes a general computational framework for solving nonconvex optimisation problems such as the penalized M-estimator \(\mathrm{argmin}_{\beta\in{\mathbb R}^d}\{ {\mathcal L}(\beta) + {\mathcal R}_{\lambda}(\beta)\}\), where \({\mathcal L}(\beta)\) is a smooth loss function, \({\mathcal R}_{\lambda}(\beta)\) is a sparsity-inducing penalty with a regularization parameter \(\lambda\). The proposed strategy enables the simultaneous control of the algorithmic complexity and the statistical error when fitting high-dimensional models appearing in various problems including low rank matrix completion problems, high-dimensional graphical models and quantile regression.
    0 references
    algorithmic statistics
    0 references
    iteration complexity
    0 references
    local adaptive MM
    0 references
    nonconvex statistical optimization
    0 references
    optimal rate of convergence
    0 references
    0 references
    0 references
    0 references
    0 references

    Identifiers

    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references