Real-time reinforcement learning by sequential actor-critics and experience replay
From MaRDI portal
Publication:1784532
Recommendations
Cites work
- scientific article; zbMATH DE number 1804129 (Why is no real title available?)
- scientific article; zbMATH DE number 3841285 (Why is no real title available?)
- scientific article; zbMATH DE number 1043533 (Why is no real title available?)
- AN ANALYSIS OF EXPERIENCE REPLAY IN TEMPORAL DIFFERENCE LEARNING
- An analysis of temporal-difference learning with function approximation
- Efficient Dynamic Computer Simulation of Robotic Mechanisms
- Least squares policy evaluation algorithms with linear function approximation
- Natural actor-critic algorithms
- OnActor-Critic Algorithms
- \({\mathcal Q}\)-learning
Cited in
(13)- Integral reinforcement learning and experience replay for adaptive optimal control of partially-unknown constrained-input continuous-time systems
- Experience selection in deep reinforcement learning for control
- A Small Gain Analysis of Single Timescale Actor Critic
- TD-regularized actor-critic methods
- Safe adaptive output-feedback optimal control of a class of linear systems
- Actor prioritized experience replay
- Deep reinforcement learning via good choice resampling experience replay memory
- Recursive estimation in piecewise affine systems using parameter identifiers and concurrent learning
- Efficient sample reuse in policy gradients with parameter-based exploration
- Artificial Intelligence and Soft Computing - ICAISC 2004
- Reward-weighted regression with sample reuse for direct policy search in reinforcement learning
- A Lie group PMP approach for optimal stabilization and tracking control of autonomous underwater vehicles
- Autonomous reinforcement learning with experience replay
This page was built for publication: Real-time reinforcement learning by sequential actor-critics and experience replay
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q1784532)