A Simultaneous Perturbation Stochastic Approximation-Based Actor–Critic Algorithm for Markov Decision Processes
From MaRDI portal
Publication:5273731
DOI10.1109/TAC.2004.825622zbMath1365.90254OpenAlexW2063123975MaRDI QIDQ5273731
Shalabh Bhatnagar, Shishir Kumar
Publication date: 12 July 2017
Published in: IEEE Transactions on Automatic Control (Search for Journal in Brave)
Full work available at URL: https://doi.org/10.1109/tac.2004.825622
Related Items (6)
Multiscale Q-learning with linear function approximation ⋮ Actor-critic algorithms for hierarchical Markov decision processes ⋮ Reinforcement learning based algorithms for average cost Markov decision processes ⋮ Approximate stochastic annealing for online control of infinite horizon Markov decision processes ⋮ Smoothed functional-based gradient algorithms for off-policy reinforcement learning: a non-asymptotic viewpoint ⋮ Natural actor-critic algorithms
This page was built for publication: A Simultaneous Perturbation Stochastic Approximation-Based Actor–Critic Algorithm for Markov Decision Processes