Nonlinear Gradient Mappings and Stochastic Optimization: A General Framework with Applications to Heavy-Tail Noise
DOI10.1137/21M145896XzbMATH Open1519.90136arXiv2204.02593MaRDI QIDQ6155875FDOQ6155875
Dušan Jakovetić, Author name not available (Why is that?), Anit Kumar Sahu, Dragana Bajović, Author name not available (Why is that?), Soummya Kar
Publication date: 7 June 2023
Published in: SIAM Journal on Optimization (Search for Journal in Brave)
Full work available at URL: https://arxiv.org/abs/2204.02593
asymptotic normalitystochastic approximationstochastic optimizationconvergence ratenonlinear mappingstochastic gradient descentmean square analysisheavy-tail noise
Numerical mathematical programming methods (65K05) Learning and adaptive systems in artificial intelligence (68T05) Convex programming (90C25) Stochastic approximation (62L20) Stochastic programming (90C15)
Cites Work
- Title not available (Why is that?)
- On the Use of Stochastic Hessian Information in Optimization Methods for Machine Learning
- Robust Stochastic Approximation Approach to Stochastic Programming
- On stochastic gradient and subgradient methods with adaptive steplength sequences
- Optimal Stochastic Approximation Algorithms for Strongly Convex Stochastic Composite Optimization I: A Generic Algorithmic Framework
- Optimal Stochastic Approximation Algorithms for Strongly Convex Stochastic Composite Optimization, II: Shrinking Procedures and Optimal Algorithms
- Lectures on stochastic programming. Modeling and theory.
- Stable laws and domains of attraction in free probability theory
- Distributed Parameter Estimation in Sensor Networks: Nonlinear Observation Models and Imperfect Communication
- Large-Scale Machine Learning with Stochastic Gradient Descent
- Adaptive estimation algorithms (convergence, optimality, stability)
- Algorithms of robust stochastic optimization based on mirror descent method
- Optimization Methods for Large-Scale Machine Learning
- Robust Consensus in the Presence of Impulsive Channel Noise
- Perturbed Iterate Analysis for Asynchronous Stochastic Optimization
- On the Adaptivity of Stochastic Gradient-Based Optimization
- Stochastic distributed learning with gradient quantization and double-variance reduction
Cited In (3)
- Nonlinear Consensus+Innovations under Correlated Heavy-Tailed Noises: Mean Square Convergence Rate and Asymptotics
- Cauchy noise loss for stochastic optimization of random matrix models via free deterministic equivalents
- Stochastic generalized gradient methods for training nonconvex nonsmooth neural networks
This page was built for publication: Nonlinear Gradient Mappings and Stochastic Optimization: A General Framework with Applications to Heavy-Tail Noise
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6155875)