I-LAMM for sparse learning: simultaneous control of algorithmic complexity and statistical error (Q1750288): Difference between revisions

From MaRDI portal
Import240304020342 (talk | contribs)
Set profile property.
Importer (talk | contribs)
Changed an Item
Property / arXiv ID
 
Property / arXiv ID: 1507.01037 / rank
 
Normal rank

Revision as of 21:39, 18 April 2024

scientific article
Language Label Description Also known as
English
I-LAMM for sparse learning: simultaneous control of algorithmic complexity and statistical error
scientific article

    Statements

    I-LAMM for sparse learning: simultaneous control of algorithmic complexity and statistical error (English)
    0 references
    0 references
    0 references
    0 references
    0 references
    18 May 2018
    0 references
    This paper proposes a general computational framework for solving nonconvex optimisation problems such as the penalized M-estimator \(\mathrm{argmin}_{\beta\in{\mathbb R}^d}\{ {\mathcal L}(\beta) + {\mathcal R}_{\lambda}(\beta)\}\), where \({\mathcal L}(\beta)\) is a smooth loss function, \({\mathcal R}_{\lambda}(\beta)\) is a sparsity-inducing penalty with a regularization parameter \(\lambda\). The proposed strategy enables the simultaneous control of the algorithmic complexity and the statistical error when fitting high-dimensional models appearing in various problems including low rank matrix completion problems, high-dimensional graphical models and quantile regression.
    0 references
    algorithmic statistics
    0 references
    iteration complexity
    0 references
    local adaptive MM
    0 references
    nonconvex statistical optimization
    0 references
    optimal rate of convergence
    0 references

    Identifiers

    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references