Stochastic compositional gradient descent: algorithms for minimizing compositions of expected-value functions
From MaRDI portal
Abstract: Classical stochastic gradient methods are well suited for minimizing expected-value objective functions. However, they do not apply to the minimization of a nonlinear function involving expected values or a composition of two expected-value functions, i.e., problems of the form . In order to solve this stochastic composition problem, we propose a class of stochastic compositional gradient descent (SCGD) algorithms that can be viewed as stochastic versions of quasi-gradient method. SCGD update the solutions based on noisy sample gradients of and use an auxiliary variable to track the unknown quantity . We prove that the SCGD converge almost surely to an optimal solution for convex optimization problems, as long as such a solution exists. The convergence involves the interplay of two iterations with different time scales. For nonsmooth convex problems, the SCGD achieve a convergence rate of in the general case and in the strongly convex case, after taking samples. For smooth convex problems, the SCGD can be accelerated to converge at a rate of in the general case and in the strongly convex case. For nonconvex problems, we prove that any limit point generated by SCGD is a stationary point, for which we also provide the convergence rate analysis. Indeed, the stochastic setting where one wants to optimize compositions of expected-value functions is very common in practice. The proposed SCGD methods find wide applications in learning, estimation, dynamic programming, etc.
Recommendations
- Accelerating Stochastic Composition Optimization
- Multilevel stochastic gradient methods for nested composition optimization
- A Stochastic Subgradient Method for Nonsmooth Nonconvex Multilevel Composition Optimization
- A smoothing stochastic gradient method for composite optimization
- Stochastic model-based minimization of weakly convex functions
Cites work
- scientific article; zbMATH DE number 3850830 (Why is no real title available?)
- scientific article; zbMATH DE number 3790208 (Why is no real title available?)
- scientific article; zbMATH DE number 48727 (Why is no real title available?)
- scientific article; zbMATH DE number 51132 (Why is no real title available?)
- scientific article; zbMATH DE number 193918 (Why is no real title available?)
- scientific article; zbMATH DE number 1972910 (Why is no real title available?)
- scientific article; zbMATH DE number 3449561 (Why is no real title available?)
- scientific article; zbMATH DE number 845714 (Why is no real title available?)
- A two Timescale Stochastic Approximation Scheme for Simulation-Based Parametric Optimization
- Acceleration of Stochastic Approximation by Averaging
- Convergence rate of linear two-time-scale stochastic approximation.
- Distributed asynchronous incremental subgradient methods
- High-dimensional analysis of semidefinite relaxations for sparse principal components
- Incremental constraint projection methods for variational inequalities
- Incremental proximal methods for large scale convex optimization
- Incremental subgradient methods for nondifferentiable optimization
- Information-Theoretic Lower Bounds on the Oracle Complexity of Stochastic Convex Optimization
- Lectures on Stochastic Programming
- Model selection and estimation in the Gaussian graphical model
- On Upper Functions for Stochastic Approximation Procedures
- Optimal Stochastic Approximation Algorithms for Strongly Convex Stochastic Composite Optimization I: A Generic Algorithmic Framework
- Optimal stochastic approximation algorithms for strongly convex stochastic composite optimization. II: Shrinking procedures and optimal algorithms
- Random algorithms for convex minimization problems
- Sparse additive models
- Statistical estimation of composite risk functionals and risk optimization problems
- Stochastic Estimation of the Maximum of a Regression Function
- Stochastic approximation with two time scales
- Stochastic approximation. A dynamical systems viewpoint.
- Variable selection in nonparametric additive models
Cited in
(38)- Momentum-based accelerated mirror descent stochastic approximation for robust topology optimization under stochastic loads
- A Two-Timescale Stochastic Algorithm Framework for Bilevel Optimization: Complexity Analysis and Application to Actor-Critic
- A Stochastic Subgradient Method for Nonsmooth Nonconvex Multilevel Composition Optimization
- Stochastic multilevel composition optimization algorithms with level-independent convergence rates
- Accelerating Stochastic Composition Optimization
- Stochastic Methods for Composite and Weakly Convex Optimization Problems
- On the information-adaptive variants of the ADMM: an iteration complexity perspective
- Gaussian fuzzy theoretic analysis for variational learning of nested compositions
- Solving Nonsmooth and Nonconvex Compound Stochastic Programs with Applications to Risk Measure Minimization
- A Stochastic Composite Augmented Lagrangian Method for Reinforcement Learning
- On the sample complexity of actor-critic method for reinforcement learning with function approximation
- Zeroth-order stochastic compositional algorithms for risk-aware learning
- Stochastic model-based minimization of weakly convex functions
- Dynamic stochastic approximation for multi-stage stochastic optimization
- Hybrid SGD algorithms to solve stochastic composite optimization problems with application in sparse portfolio selection problems
- scientific article; zbMATH DE number 7370578 (Why is no real title available?)
- Distributed stochastic compositional optimization problems over directed networks
- Multilevel stochastic gradient methods for nested composition optimization
- Proximally guided stochastic subgradient method for nonsmooth, nonconvex problems
- A convergence analysis of the perturbed compositional gradient flow: averaging principle and normal deviations
- A Single Timescale Stochastic Approximation Method for Nested Stochastic Optimization
- A functional model method for nonconvex nonsmooth conditional stochastic optimization
- Stochastic nested primal-dual method for nonconvex constrained composition optimization
- A stochastic subgradient method for distributionally robust non-convex and non-smooth learning
- scientific article; zbMATH DE number 7415103 (Why is no real title available?)
- A single timescale stochastic quasi-Newton method for stochastic optimization
- A Two-Time-Scale Stochastic Optimization Framework with Applications in Control and Reinforcement Learning
- Sample complexity of sample average approximation for conditional stochastic optimization
- Probability maximization via Minkowski functionals: convex representations and tractable resolution
- A hybrid stochastic optimization framework for composite nonconvex optimization
- Stochastic composition optimization of functions without Lipschitz continuous gradient
- Finite-time analysis and restarting scheme for linear two-time-scale stochastic approximation
- Multilevel composite stochastic optimization via nested variance reduction
- Stochastic variance-reduced prox-linear algorithms for nonconvex composite optimization
- Streaming constrained binary logistic regression with online standardized data
- Adaptive primal-dual stochastic gradient method for expectation-constrained convex stochastic programs
- Primal-Dual Stochastic Gradient Method for Convex Programs with Many Functional Constraints
- Near-optimal stochastic approximation for online principal component estimation
This page was built for publication: Stochastic compositional gradient descent: algorithms for minimizing compositions of expected-value functions
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q507334)