Optimal Rates for Zero-Order Convex Optimization: The Power of Two Function Evaluations

From MaRDI portal
Revision as of 14:29, 6 March 2024 by Import240305080351 (talk | contribs) (Created automatically from import240305080351)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)

Publication:2978646

DOI10.1109/TIT.2015.2409256zbMath1359.90155arXiv1312.2139OpenAlexW2171830216MaRDI QIDQ2978646

Andre Wibisono, Michael I. Jordan, Martin J. Wainwright, John C. Duchi

Publication date: 28 April 2017

Published in: IEEE Transactions on Information Theory (Search for Journal in Brave)

Abstract: We consider derivative-free algorithms for stochastic and non-stochastic convex optimization problems that use only function values rather than gradients. Focusing on non-asymptotic bounds on convergence rates, we show that if pairs of function values are available, algorithms for $d$-dimensional optimization that use gradient estimates based on random perturbations suffer a factor of at most $sqrt{d}$ in convergence rate over traditional stochastic gradient methods. We establish such results for both smooth and non-smooth cases, sharpening previous analyses that suggested a worse dimension dependence, and extend our results to the case of multiple ($m ge 2$) evaluations. We complement our algorithmic development with information-theoretic lower bounds on the minimax convergence rate of such problems, establishing the sharpness of our achievable results up to constant (sometimes logarithmic) factors.


Full work available at URL: https://arxiv.org/abs/1312.2139






Related Items (58)

Full-low evaluation methods for derivative-free optimizationGradient-free two-point methods for solving stochastic nonsmooth convex optimization problems with small non-random noisesStochastic zeroth-order discretizations of Langevin diffusions for Bayesian inferenceNew First-Order Algorithms for Stochastic Variational InequalitiesDerivative-Free Methods for Policy Optimization: Guarantees for Linear Quadratic SystemsA theoretical and empirical comparison of gradient approximations in derivative-free optimizationFinite Difference Gradient Approximation: To Randomize or Not?Zeroth-order algorithms for stochastic distributed nonconvex optimizationZeroth-Order Stochastic Compositional Algorithms for Risk-Aware LearningStochastic online optimization. Single-point and multi-point non-linear multi-armed bandits. Convex and strongly-convex caseGradient-free distributed optimization with exact convergenceAn Accelerated Method for Derivative-Free Smooth Stochastic Convex OptimizationDecentralized online convex optimization based on signs of relative statesPersonalized optimization with user's feedbackImproved complexities for stochastic conditional gradient methods under interpolation-like conditionsOn the convergence rate issues of general Markov search for global minimumZeroth-order optimization with orthogonal random directionsGradient-free federated learning methods with \(l_1\) and \(l_2\)-randomization for non-smooth convex stochastic optimization problemsGradient-free methods for non-smooth convex stochastic optimization with heavy-tailed noise on convex compactNon-smooth setting of stochastic decentralized convex optimization problem over time-varying graphsA Zeroth-Order Proximal Stochastic Gradient Method for Weakly Convex Stochastic OptimizationA gradient‐free distributed optimization method for convex sum of nonconvex cost functionsDirect Search Based on Probabilistic Descent in Reduced SpacesDistributed Nash equilibrium learning: A second‐order proximal algorithmOptimistic optimisation of composite objective with exponentiated updateZeroth-order feedback optimization for cooperative multi-agent systemsFederated learning for minimizing nonsmooth convex loss functionsRe-thinking high-dimensional mathematical statistics. Abstracts from the workshop held May 15--21, 2022Unnamed ItemExact optimization: Part IAdaptive sampling quasi-Newton methods for zeroth-order stochastic optimizationOnline distributed dual averaging algorithm for multi-agent bandit optimization over time-varying general directed networksUnifying framework for accelerated randomized methods in convex optimizationZeroth-order nonconvex stochastic optimization: handling constraints, high dimensionality, and saddle pointsAdaptive Catalyst for Smooth Convex OptimizationOnline Statistical Inference for Stochastic Optimization via Kiefer-Wolfowitz MethodsConvergence guarantees for forward gradient descent in the linear regression modelExpected decrease for derivative-free algorithms using random subspacesUnnamed ItemAn accelerated directional derivative method for smooth stochastic convex optimizationSmall errors in random zeroth-order optimization are imaginaryOn the upper bound for the expectation of the norm of a vector uniformly distributed on the sphere and the phenomenon of concentration of uniform measure on the sphereWorst case complexity of direct search under convexityDistributed zeroth-order optimization: convergence rates that match centralized counterpartStochastic adversarial noise in the ``black box optimization problemAccelerated zero-order SGD method for solving the black box optimization problem under ``overparametrization conditionA stochastic subspace approach to gradient-free optimization in high dimensionsStochastic zeroth order descent with structured directionsGradient-Free Methods with Inexact Oracle for Convex-Concave Stochastic Saddle-Point ProblemA zeroth order method for stochastic weakly convex optimizationA new one-point residual-feedback oracle for black-box learning and controlDerivative-free optimization methodsDistributed Subgradient-Free Stochastic Optimization Algorithm for Nonsmooth Convex Functions over Time-Varying NetworksModel-free linear quadratic regulatorUnnamed ItemDistributed online bandit optimization under random quantizationNoisy zeroth-order optimization for non-smooth saddle point problemsOne-point gradient-free methods for smooth and non-smooth saddle-point problems





This page was built for publication: Optimal Rates for Zero-Order Convex Optimization: The Power of Two Function Evaluations