Extreme state aggregation beyond Markov decision processes
From MaRDI portal
Publication:329613
DOI10.1016/J.TCS.2016.07.032zbMATH Open1353.68236OpenAlexW2482498454WikidataQ58012178 ScholiaQ58012178MaRDI QIDQ329613FDOQ329613
Publication date: 21 October 2016
Published in: Theoretical Computer Science (Search for Journal in Brave)
Full work available at URL: https://doi.org/10.1016/j.tcs.2016.07.032
Recommendations
- Extreme state aggregation beyond MDPs
- Selecting near-optimal approximate state representations in reinforcement learning
- Adaptive aggregation for reinforcement learning in average reward Markov decision processes
- Reinforcement learning of non-Markov decision processes
- Relative value iteration algorithm with soft state aggregation
Learning and adaptive systems in artificial intelligence (68T05) Markov and semi-Markov decision processes (90C40)
Cites Work
- Title not available (Why is that?)
- Title not available (Why is that?)
- Equivalence notions and model minimization in Markov decision processes
- A General Approach to the Strong Law of Large Numbers
- Universal artificial intelligence. Sequential decisions based on algorithmic probability.
- On a general approach to the strong laws of large numbers
- Reinforcement learning in finite MDPs: PAC analysis
- Near-optimal regret bounds for reinforcement learning
- Extreme State Aggregation beyond MDPs
- Title not available (Why is that?)
- Pseudometrics for State Aggregation in Average Reward Markov Decision Processes
- General time consistent discounting
- Near-optimal PAC bounds for discounted MDPs
- Consistency of Feature Markov Processes
- Learning Theory and Kernel Machines
Cited In (3)
This page was built for publication: Extreme state aggregation beyond Markov decision processes
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q329613)