Bayesian optimistic Kullback-Leibler exploration
DOI10.1007/S10994-018-5767-4zbMATH Open1493.68304OpenAlexW2903743315WikidataQ128709924 ScholiaQ128709924MaRDI QIDQ2425228FDOQ2425228
Authors: Kanghoon Lee, Geon-Hyeong Kim, Daniel D. Lee, Kee-Eung Kim, Pedro A. Ortega
Publication date: 26 June 2019
Published in: Machine Learning (Search for Journal in Brave)
Full work available at URL: https://doi.org/10.1007/s10994-018-5767-4
Recommendations
- Bayesian Reinforcement Learning with Exploration
- Scalable and efficient Bayes-adaptive reinforcement learning based on Monte-Carlo tree search
- Exploration in relational domains for model-based reinforcement learning
- Dual control for approximate Bayesian reinforcement learning
- Using trajectory data to improve Bayesian optimization for reinforcement learning
Statistical aspects of information-theoretic topics (62B10) Bayesian inference (62F15) Learning and adaptive systems in artificial intelligence (68T05)
Cites Work
- 10.1162/153244303765208377
- Title not available (Why is that?)
- Exploration-exploitation tradeoff using variance estimates in multi-armed bandits
- Title not available (Why is that?)
- Near-optimal regret bounds for reinforcement learning
- An analysis of model-based interval estimation for Markov decision processes
- Near-optimal reinforcement learning in polynomial time
Cited In (6)
- Bayesian Reinforcement Learning with Exploration
- Exploration in relational domains for model-based reinforcement learning
- Using trajectory data to improve Bayesian optimization for reinforcement learning
- Title not available (Why is that?)
- Dual control for approximate Bayesian reinforcement learning
- Optimistic reinforcement learning by forward Kullback-Leibler divergence optimization
This page was built for publication: Bayesian optimistic Kullback-Leibler exploration
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q2425228)