Optimal primal-dual methods for a class of saddle point problems

From MaRDI portal
Publication:5245366

DOI10.1137/130919362zbMATH Open1329.90090arXiv1309.5548OpenAlexW2092851214MaRDI QIDQ5245366FDOQ5245366

Yuyuan Ouyang, Yunmei Chen, Guanghui Lan

Publication date: 8 April 2015

Published in: SIAM Journal on Optimization (Search for Journal in Brave)

Abstract: We present a novel accelerated primal-dual (APD) method for solving a class of deterministic and stochastic saddle point problems (SPP). The basic idea of this algorithm is to incorporate a multi-step acceleration scheme into the primal-dual method without smoothing the objective function. For deterministic SPP, the APD method achieves the same optimal rate of convergence as Nesterov's smoothing technique. Our stochastic APD method exhibits an optimal rate of convergence for stochastic SPP not only in terms of its dependence on the number of the iteration, but also on a variety of problem parameters. To the best of our knowledge, this is the first time that such an optimal algorithm has been developed for stochastic SPP in the literature. Furthermore, for both deterministic and stochastic SPP, the developed APD algorithms can deal with the situation when the feasible region is unbounded, as long as a saddle point exists. In the unbounded case, we incorporate the modified termination criterion introduced by Monteiro and Svaiter in solving SPP problem posed as monotone inclusion, and demonstrate that the rate of convergence of the APD method depends on the distance from the initial point to the set of optimal solutions.


Full work available at URL: https://arxiv.org/abs/1309.5548




Recommendations





Cited In (87)





This page was built for publication: Optimal primal-dual methods for a class of saddle point problems

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q5245366)