Pure exploration in finitely-armed and continuous-armed bandits
From MaRDI portal
Publication:2431430
DOI10.1016/j.tcs.2010.12.059zbMath1214.62082OpenAlexW2108794978MaRDI QIDQ2431430
Rémi Munos, Sébastien Bubeck, Gilles Stoltz
Publication date: 14 April 2011
Published in: Theoretical Computer Science (Search for Journal in Brave)
Full work available at URL: https://doi.org/10.1016/j.tcs.2010.12.059
Related Items (16)
Bandit Theory: Applications to Learning Healthcare Systems and Clinical Trials ⋮ Robust Learning of Consumer Preferences ⋮ Intrinsically motivated model learning for developing curious robots ⋮ Information theory for ranking and selection ⋮ Unnamed Item ⋮ Simple and cumulative regret for continuous noisy optimization ⋮ Learning the distribution with largest mean: two bandit frameworks ⋮ Gaussian process bandits with adaptive discretization ⋮ Deep learning for ranking response surfaces with applications to optimal stopping problems ⋮ Unnamed Item ⋮ A bad arm existence checking problem: how to utilize asymmetric problem structure? ⋮ Bayesian Incentive-Compatible Bandit Exploration ⋮ A PAC algorithm in relative precision for bandit problem with costly sampling ⋮ An asymptotically optimal strategy for constrained multi-armed bandit problems ⋮ Trading utility and uncertainty: applying the value of information to resolve the exploration-exploitation dilemma in reinforcement learning ⋮ Sequential Design for Ranking Response Surfaces
Cites Work
- Exploration-exploitation tradeoff using variance estimates in multi-armed bandits
- Asymptotically efficient adaptive allocation rules
- Learning Theory
- The Nonstochastic Multiarmed Bandit Problem
- Probability Inequalities for Sums of Bounded Random Variables
- Some aspects of the sequential design of experiments
- Combinatorial methods in density estimation
- Finite-time analysis of the multiarmed bandit problem
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
This page was built for publication: Pure exploration in finitely-armed and continuous-armed bandits