Bellman's principle of optimality and deep reinforcement learning for time-varying tasks
From MaRDI portal
Publication:5043501
DOI10.1080/00207179.2021.1913516zbMATH Open1500.93144OpenAlexW3146773041MaRDI QIDQ5043501FDOQ5043501
Authors: Alessandro Giuseppi, Antonio Pietrabissa Error creating thumbnail:
Publication date: 6 October 2022
Published in: International Journal of Control (Search for Journal in Brave)
Full work available at URL: https://doi.org/10.1080/00207179.2021.1913516
Recommendations
- On Generalized Bellman Equations and Temporal-Difference Learning
- On Generalized Bellman Equations and Temporal-Difference Learning
- Reinforcement Learning and Stochastic Optimization
- The Bellman's principle of optimality in the discounted dynamic programming
- From reinforcement learning to optimal control: a unified framework for sequential decisions
- Time-varying policy rule under learning
- Approximate dynamic programming via iterated Bellman inequalities
- Policy iterations for reinforcement learning problems in continuous time and space -- fundamental theory and methods
Cites Work
- Title not available (Why is that?)
- Dynamic programming and optimal control. Vol. 1.
- Title not available (Why is that?)
- An iterative adaptive dynamic programming method for solving a class of nonlinear zero-sum differential games
- Modelling and solving resource allocation problems via a dynamic programming approach
- Title not available (Why is that?)
- Allocating resources via price management systems: a dynamic programming-based approach
- Adaptive dynamic programming for discrete-time linear quadratic regulation based on multirate generalised policy iteration
Cited In (3)
Uses Software
This page was built for publication: Bellman's principle of optimality and deep reinforcement learning for time-varying tasks
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q5043501)