A proximal subgradient algorithm with extrapolation for structured nonconvex nonsmooth problems (Q6141533)
From MaRDI portal
scientific article; zbMATH DE number 7780865
Language | Label | Description | Also known as |
---|---|---|---|
English | A proximal subgradient algorithm with extrapolation for structured nonconvex nonsmooth problems |
scientific article; zbMATH DE number 7780865 |
Statements
A proximal subgradient algorithm with extrapolation for structured nonconvex nonsmooth problems (English)
0 references
19 December 2023
0 references
In the paper under review, the authors study a following broad optimization problem which has many important applications in diverse areas, including power control problems, compressed sensing, portfolio optimization, supply chain problems, image segmentation and many others. The objective function is formed by the sum of a possibly nonsmooth nonconvex function and a differentiable function with Lipschitz continuous gradient, subtracted by a weakly convex function. This general framework allows for problems involving nonconvex loss functions and problems with specific nonconvex constraints. The problem is: \(\min_{x\in C} F(x)\) where \(F(x) := f (x) + h(Ax)-g(x)\) and where \(C\) is a nonempty closed subset of a finite-dimensional real Hilbert space \(H\), \(A\) is a linear mapping from \(H\) to another finite-dimensional real Hilbert space, \(f : H\to (-\infty,\infty]\) is a proper lower semicontinuous (possibly nonsmooth and nonconvex) function, \(h\) is a real valued differentiable (possibly nonconvex) function whose gradient is Lipschitz continuous and \(g:H\to (-\infty,\infty]\) is a continuous weakly convex function with modulus on an open convex set containing \(C\). Two examples are: (1) From statistical learning, \(\min_{x\in \mathbb R^d}(\phi(x) + \gamma r(x))\) where \(\phi\) is called a loss function which measures the data fitting, \(r\) is a regularization which promotes specific structure in the solution such as sparsity, and \(\gamma>0\) is a weighting parameter. (2) \(\min_{x\in \mathbb R^d} (f (x) + h(x)- g(x))\). The paper is well written with a good set of references.
0 references
composite optimization problem
0 references
difference of convex
0 references
distributed energy resources
0 references
extrapolation
0 references
optimal power flow
0 references
proximal subgradient algorithm
0 references
0 references
0 references
0 references
0 references
0 references
0 references
0 references
0 references
0 references
0 references