Near-optimal PAC bounds for discounted MDPs
From MaRDI portal
Publication:465258
DOI10.1016/J.TCS.2014.09.029zbMATH Open1360.68528OpenAlexW1969276875WikidataQ58012230 ScholiaQ58012230MaRDI QIDQ465258FDOQ465258
Authors: Tor Lattimore, Marcus Hutter
Publication date: 31 October 2014
Published in: Theoretical Computer Science (Search for Journal in Brave)
Full work available at URL: https://doi.org/10.1016/j.tcs.2014.09.029
Recommendations
Cites Work
- Asymptotically efficient adaptive allocation rules
- Concentration Inequalities and Martingale Inequalities: A Survey
- The sample complexity of exploration in the multi-armed bandit problem
- The variance of discounted Markov decision processes
- Reinforcement learning in finite MDPs: PAC analysis
- Near-optimal regret bounds for reinforcement learning
- An analysis of model-based interval estimation for Markov decision processes
- PAC Bounds for Discounted MDPs
- Minimax PAC bounds on the sample complexity of reinforcement learning with a generative model
- Bayesian Reinforcement Learning with Exploration
Cited In (10)
- Minimax PAC bounds on the sample complexity of reinforcement learning with a generative model
- Reinforcement learning in finite MDPs: PAC analysis
- Near-optimal regret bounds for reinforcement learning
- Complexity bounds for approximately solving discounted MDPs by value iterations
- Title not available (Why is that?)
- Near-optimal reinforcement learning in polynomial time
- Extreme state aggregation beyond Markov decision processes
- Is Temporal Difference Learning Optimal? An Instance-Dependent Analysis
- Optimistic Posterior Sampling for Reinforcement Learning: Worst-Case Regret Bounds
- PAC Bounds for Discounted MDPs
This page was built for publication: Near-optimal PAC bounds for discounted MDPs
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q465258)