A stochastic subgradient method for distributionally robust non-convex and non-smooth learning

From MaRDI portal
Publication:2159458

DOI10.1007/S10957-022-02063-6zbMATH Open1492.90103arXiv2006.04873OpenAlexW4284970065MaRDI QIDQ2159458FDOQ2159458

Landi Zhu, Mert Gürbüzbalaban, Andrzej Ruszczyński

Publication date: 1 August 2022

Published in: Journal of Optimization Theory and Applications (Search for Journal in Brave)

Abstract: We consider a distributionally robust formulation of stochastic optimization problems arising in statistical learning, where robustness is with respect to uncertainty in the underlying data distribution. Our formulation builds on risk-averse optimization techniques and the theory of coherent risk measures. It uses semi-deviation risk for quantifying uncertainty, allowing us to compute solutions that are robust against perturbations in the population data distribution. We consider a large family of loss functions that can be non-convex and non-smooth and develop an efficient stochastic subgradient method. We prove that it converges to a point satisfying the optimality conditions. To our knowledge, this is the first method with rigorous convergence guarantees in the context of non-convex non-smooth distributionally robust stochastic optimization. Our method can achieve any desired level of robustness with little extra computational cost compared to population risk minimization. We also illustrate the performance of our algorithm on real datasets arising in convex and non-convex supervised learning problems.


Full work available at URL: https://arxiv.org/abs/2006.04873




Recommendations




Cites Work


Cited In (3)

Uses Software





This page was built for publication: A stochastic subgradient method for distributionally robust non-convex and non-smooth learning

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q2159458)