I-LAMM for sparse learning: simultaneous control of algorithmic complexity and statistical error (Q1750288): Difference between revisions
From MaRDI portal
ReferenceBot (talk | contribs) Changed an Item |
Set OpenAlex properties. |
||
Property / OpenAlex ID | |||
Property / OpenAlex ID: W2964248738 / rank | |||
Normal rank |
Latest revision as of 10:24, 30 July 2024
scientific article
Language | Label | Description | Also known as |
---|---|---|---|
English | I-LAMM for sparse learning: simultaneous control of algorithmic complexity and statistical error |
scientific article |
Statements
I-LAMM for sparse learning: simultaneous control of algorithmic complexity and statistical error (English)
0 references
18 May 2018
0 references
This paper proposes a general computational framework for solving nonconvex optimisation problems such as the penalized M-estimator \(\mathrm{argmin}_{\beta\in{\mathbb R}^d}\{ {\mathcal L}(\beta) + {\mathcal R}_{\lambda}(\beta)\}\), where \({\mathcal L}(\beta)\) is a smooth loss function, \({\mathcal R}_{\lambda}(\beta)\) is a sparsity-inducing penalty with a regularization parameter \(\lambda\). The proposed strategy enables the simultaneous control of the algorithmic complexity and the statistical error when fitting high-dimensional models appearing in various problems including low rank matrix completion problems, high-dimensional graphical models and quantile regression.
0 references
algorithmic statistics
0 references
iteration complexity
0 references
local adaptive MM
0 references
nonconvex statistical optimization
0 references
optimal rate of convergence
0 references
0 references
0 references
0 references
0 references
0 references
0 references
0 references