A derivative-free comirror algorithm for convex optimization

From MaRDI portal
Revision as of 21:32, 4 February 2024 by Import240129110113 (talk | contribs) (Created automatically from import240129110113)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)

Publication:3458813

DOI10.1080/10556788.2014.968158zbMath1328.90104arXiv1210.6403OpenAlexW2162546528MaRDI QIDQ3458813

Walaa M. Moursi, Heinz H. Bauschke, Warren L. Hare

Publication date: 28 December 2015

Published in: Unnamed Author (Search for Journal in Brave)

Abstract: We consider $min{f(x):g(x) le 0, ~xin X},$ where $X$ is a compact convex subset of $RR^m$, and $f$ and $g$ are continuous convex functions defined on an open neighbourhood of $X$. We work in the setting of derivative-free optimization, assuming that $f$ and $g$ are available through a black-box that provides only function values for a lower-$mathcal{C}^2$ representation of the functions. We present a derivative-free optimization variant of the $eps$-comirror algorithm cite{BBTGBT2010}. Algorithmic convergence hinges on the ability to accurately approximate subgradients of lower-$mathcal{C}^2$ functions, which we prove is possible through linear interpolation. We provide convergence analysis that quantifies the difference between the function values of the iterates and the optimal function value. We find that the DFO algorithm we develop has the same convergence result as the original gradient-based algorithm. We present some numerical testing that demonstrate the practical feasibility of the algorithm, and conclude with some directions for further research.


Full work available at URL: https://arxiv.org/abs/1210.6403





Cites Work


Related Items (4)





This page was built for publication: A derivative-free comirror algorithm for convex optimization