Approximate policy iteration for Markov decision processes via quantitative adaptive aggregations
From MaRDI portal
Publication:1990495
DOI10.1007/978-3-319-46520-3_2zbMATH Open1398.90200OpenAlexW2523642882MaRDI QIDQ1990495FDOQ1990495
Milan Češka, Alessandro Abate, Marta Kwiatkowska
Publication date: 25 October 2018
Full work available at URL: https://doi.org/10.1007/978-3-319-46520-3_2
Recommendations
- scientific article; zbMATH DE number 3871063
- Approximation of infinite horizon discounted cost Markov decision processes
- A time aggregation approach to Markov decision processes
- On the asymptotic optimality of finite approximations to Markov decision processes with Borel spaces
- Approximation of Markov decision processes with general state space
Cited In (7)
- Optimal learning with \textit{Q}-aggregation
- Aggregation of the policy iteration method for nearly completely decomposable Markov chains
- Relative value iteration algorithm with soft state aggregation
- Introduction to internally consistent modeling, aggregation, inference, and policy
- Adaptive policy-iteration and policy-value-iteration for discounted Markov decision processes
- On State Aggregation to Approximate Complex Value Functions in Large-Scale Markov Decision Processes
- Title not available (Why is that?)
This page was built for publication: Approximate policy iteration for Markov decision processes via quantitative adaptive aggregations
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q1990495)