An Optimal Algorithm for Bandit and Zero-Order Convex Optimization with Two-Point Feedback

From MaRDI portal
Publication:5361319

zbMath1440.90049arXiv1507.08752MaRDI QIDQ5361319

Ohad Shamir

Publication date: 27 September 2017

Full work available at URL: https://arxiv.org/abs/1507.08752




Related Items (28)

New First-Order Algorithms for Stochastic Variational InequalitiesDerivative-Free Methods for Policy Optimization: Guarantees for Linear Quadratic SystemsA theoretical and empirical comparison of gradient approximations in derivative-free optimizationFinite Difference Gradient Approximation: To Randomize or Not?Zeroth-order algorithms for stochastic distributed nonconvex optimizationAn Accelerated Method for Derivative-Free Smooth Stochastic Convex OptimizationGradient-free federated learning methods with \(l_1\) and \(l_2\)-randomization for non-smooth convex stochastic optimization problemsGradient-free methods for non-smooth convex stochastic optimization with heavy-tailed noise on convex compactNon-smooth setting of stochastic decentralized convex optimization problem over time-varying graphsSign stochastic gradient descents without bounded gradient assumption for the finite sum minimizationZeroth-order feedback optimization for cooperative multi-agent systemsRe-thinking high-dimensional mathematical statistics. Abstracts from the workshop held May 15--21, 2022Unnamed ItemAdaptive sampling quasi-Newton methods for zeroth-order stochastic optimizationUnifying framework for accelerated randomized methods in convex optimizationTechnical Note—Nonstationary Stochastic Optimization Under Lp,q-Variation MeasuresAdaptive Catalyst for Smooth Convex OptimizationRecent theoretical advances in decentralized distributed convex optimizationSmoothed functional-based gradient algorithms for off-policy reinforcement learning: a non-asymptotic viewpointAn accelerated directional derivative method for smooth stochastic convex optimizationOn the upper bound for the expectation of the norm of a vector uniformly distributed on the sphere and the phenomenon of concentration of uniform measure on the sphereGradient-Free Methods with Inexact Oracle for Convex-Concave Stochastic Saddle-Point ProblemA new one-point residual-feedback oracle for black-box learning and controlDerivative-free optimization methodsDistributed online bandit optimization under random quantizationUnnamed ItemNoisy zeroth-order optimization for non-smooth saddle point problemsOne-point gradient-free methods for smooth and non-smooth saddle-point problems




This page was built for publication: An Optimal Algorithm for Bandit and Zero-Order Convex Optimization with Two-Point Feedback