Kernel-based methods for bandit convex optimization

From MaRDI portal
Publication:4977962

DOI10.1145/3055399.3055403zbMATH Open1370.90175arXiv1607.03084OpenAlexW2963831922MaRDI QIDQ4977962FDOQ4977962


Authors: Sébastien Bubeck, Yin Tat Lee, Ronen Eldan Edit this on Wikidata


Publication date: 17 August 2017

Published in: Proceedings of the 49th Annual ACM SIGACT Symposium on Theory of Computing (Search for Journal in Brave)

Abstract: We consider the adversarial convex bandit problem and we build the first mathrmpoly(T)-time algorithm with mathrmpoly(n)sqrtT-regret for this problem. To do so we introduce three new ideas in the derivative-free optimization literature: (i) kernel methods, (ii) a generalization of Bernoulli convolutions, and (iii) a new annealing schedule for exponential weights (with increasing learning rate). The basic version of our algorithm achieves ildeO(n9.5sqrtT)-regret, and we show that a simple variant of this algorithm can be run in mathrmpoly(nlog(T))-time per step at the cost of an additional mathrmpoly(n)To(1) factor in the regret. These results improve upon the ildeO(n11sqrtT)-regret and exp(mathrmpoly(T))-time result of the first two authors, and the log(T)mathrmpoly(n)sqrtT-regret and log(T)mathrmpoly(n)-time result of Hazan and Li. Furthermore we conjecture that another variant of the algorithm could achieve ildeO(n1.5sqrtT)-regret, and moreover that this regret is unimprovable (the current best lower bound being Omega(nsqrtT) and it is achieved with linear functions). For the simpler situation of zeroth order stochastic convex optimization this corresponds to the conjecture that the optimal query complexity is of order n3/epsilon2.


Full work available at URL: https://arxiv.org/abs/1607.03084




Recommendations





Cited In (15)





This page was built for publication: Kernel-based methods for bandit convex optimization

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q4977962)