On tight bounds for function approximation error in risk-sensitive reinforcement learning
DOI10.1016/J.SYSCONLE.2021.104899zbMATH Open1478.91064OpenAlexW3149920308MaRDI QIDQ2243003FDOQ2243003
Prasenjit Karmakar, Shalabh Bhatnagar
Publication date: 10 November 2021
Published in: Systems \& Control Letters (Search for Journal in Brave)
Full work available at URL: https://doi.org/10.1016/j.sysconle.2021.104899
stochastic optimal controlstochastic systemsfunction approximationeigenvalue perturbationPerron-Frobenius eigenvaluerisk-sensitive reinforcement learning
Learning and adaptive systems in artificial intelligence (68T05) Decision theory (91B06) Optimal stochastic control (93E20)
Cites Work
- Title not available (Why is that?)
- A quantitative comparison of risk measures
- Simulation-based optimization of Markov reward processes
- Risk-averse dynamic programming for Markov decision processes
- Risk-Sensitive Optimal Control for Markov Decision Processes with Monotone Cost
- Q-Learning for Risk-Sensitive Control
- A sensitivity formula for risk-sensitive cost and the actor-critic algorithm
- Markov decision processes with average-value-at-risk criteria
- On the Maximal Eigenvector of a Positive Matrix
- Inequalities
- Comparing the Spectral Radii of Two Nonnegative Matrices
- On comparison of the Perron-Frobenius eigenvalues of two ML-matrices
- Multiplicative ergodicity and large deviations for an irreducible Markov chain.
- Stochastic approximation with `controlled Markov' noise
Cited In (3)
This page was built for publication: On tight bounds for function approximation error in risk-sensitive reinforcement learning
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q2243003)