An optimal method for stochastic composite optimization (Q431018): Difference between revisions

From MaRDI portal
Importer (talk | contribs)
Created a new Item
 
Importer (talk | contribs)
Changed an Item
Property / review text
 
The stochastic composite optimization problem under consideration consists in minimizing the sum \(f+h\) of two convex functions \(f,h\) defined on a compact convex set \(X\subset \mathbb{R}^{n}\), assuming that \(\nabla f\) and \(h\) are globally Lipschitz. The author proposes two subgradient-type methods, namely, a modified version of the mirror-descent SA method due to \textit{A. Nemirovski, A. Juditsky, G. Lan} and \textit{A. Shapiro} [SIAM J. Optim. 19, No. 4, 1574--1609 (2009; Zbl 1189.90109)], which substantially improves the rate of convergence, and an accelerated stochastic approximation method, which can achieve an optimal rate of convergence. He illustrates the advantages of the latter method over other existing methods by discussing its performance for a particular class of stochastic optimization problems.
Property / review text: The stochastic composite optimization problem under consideration consists in minimizing the sum \(f+h\) of two convex functions \(f,h\) defined on a compact convex set \(X\subset \mathbb{R}^{n}\), assuming that \(\nabla f\) and \(h\) are globally Lipschitz. The author proposes two subgradient-type methods, namely, a modified version of the mirror-descent SA method due to \textit{A. Nemirovski, A. Juditsky, G. Lan} and \textit{A. Shapiro} [SIAM J. Optim. 19, No. 4, 1574--1609 (2009; Zbl 1189.90109)], which substantially improves the rate of convergence, and an accelerated stochastic approximation method, which can achieve an optimal rate of convergence. He illustrates the advantages of the latter method over other existing methods by discussing its performance for a particular class of stochastic optimization problems. / rank
 
Normal rank
Property / reviewed by
 
Property / reviewed by: Juan-Enrique Martinez-Legaz / rank
 
Normal rank
Property / Mathematics Subject Classification ID
 
Property / Mathematics Subject Classification ID: 90C15 / rank
 
Normal rank
Property / Mathematics Subject Classification ID
 
Property / Mathematics Subject Classification ID: 90C25 / rank
 
Normal rank
Property / Mathematics Subject Classification ID
 
Property / Mathematics Subject Classification ID: 62L20 / rank
 
Normal rank
Property / zbMATH DE Number
 
Property / zbMATH DE Number: 6050450 / rank
 
Normal rank
Property / zbMATH Keywords
 
stochastic optimization
Property / zbMATH Keywords: stochastic optimization / rank
 
Normal rank
Property / zbMATH Keywords
 
optimal method
Property / zbMATH Keywords: optimal method / rank
 
Normal rank
Property / zbMATH Keywords
 
convex optimization
Property / zbMATH Keywords: convex optimization / rank
 
Normal rank
Property / zbMATH Keywords
 
complexity theory
Property / zbMATH Keywords: complexity theory / rank
 
Normal rank
Property / zbMATH Keywords
 
stochastic approximation
Property / zbMATH Keywords: stochastic approximation / rank
 
Normal rank

Revision as of 23:52, 29 June 2023

scientific article
Language Label Description Also known as
English
An optimal method for stochastic composite optimization
scientific article

    Statements

    An optimal method for stochastic composite optimization (English)
    0 references
    0 references
    26 June 2012
    0 references
    The stochastic composite optimization problem under consideration consists in minimizing the sum \(f+h\) of two convex functions \(f,h\) defined on a compact convex set \(X\subset \mathbb{R}^{n}\), assuming that \(\nabla f\) and \(h\) are globally Lipschitz. The author proposes two subgradient-type methods, namely, a modified version of the mirror-descent SA method due to \textit{A. Nemirovski, A. Juditsky, G. Lan} and \textit{A. Shapiro} [SIAM J. Optim. 19, No. 4, 1574--1609 (2009; Zbl 1189.90109)], which substantially improves the rate of convergence, and an accelerated stochastic approximation method, which can achieve an optimal rate of convergence. He illustrates the advantages of the latter method over other existing methods by discussing its performance for a particular class of stochastic optimization problems.
    0 references
    0 references
    0 references
    0 references
    0 references
    stochastic optimization
    0 references
    optimal method
    0 references
    convex optimization
    0 references
    complexity theory
    0 references
    stochastic approximation
    0 references