Modification of improved upper confidence bounds for regulating exploration in Monte-Carlo tree search
DOI10.1016/J.TCS.2016.06.034zbMATH Open1370.68263OpenAlexW2467873743MaRDI QIDQ307787FDOQ307787
Authors: Yun-Ching Liu, Yoshimasa Tsuruoka
Publication date: 5 September 2016
Published in: Theoretical Computer Science (Search for Journal in Brave)
Full work available at URL: https://doi.org/10.1016/j.tcs.2016.06.034
Recommendations
Problem solving in the context of artificial intelligence (heuristics, search strategies, etc.) (68T20) Combinatorial games (91A46) Probabilistic games; gambling (91A60)
Cites Work
- Pure exploration in multi-armed bandits problems
- Asymptotically efficient adaptive allocation rules
- Finite-time analysis of the multiarmed bandit problem
- UCB revisited: improved regret bounds for the stochastic multi-armed bandit problem
- Simple regret optimization in online planning for Markov decision processes
- Thompson sampling: an asymptotically optimal finite-time analysis
Cited In (4)
This page was built for publication: Modification of improved upper confidence bounds for regulating exploration in Monte-Carlo tree search
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q307787)