Undiscounted control policy generation for continuous-valued optimal control by approximate dynamic programming
From MaRDI portal
Publication:5043547
Abstract: We present a numerical method for generating the state-feedback control policy associated with general undiscounted, constant-setpoint, infinite-horizon, nonlinear optimal control problems with continuous state variables. The method is based on approximate dynamic programming, and is closely related to approximate policy iteration. Existing methods typically terminate based on the convergence of the control policy and either require a discounted problem formulation or demand the cost function to lie in a specific subclass of functions. The presented method extends on existing termination criteria by requiring both the control policy and the resulting system state to converge, allowing for use with undiscounted cost functions that are bounded and continuous. This paper defines the numerical method, derives the relevant underlying mathematical properties, and validates the numerical method with representative examples. A MATLAB implementation with the shown examples is freely available.
Recommendations
- Temporal difference-based policy iteration for optimal control of stochastic systems
- Controller design and value function approximation for nonlinear dynamical systems
- Optimal control for nonlinear continuous systems by adaptive dynamic programming based on fuzzy basis functions
- Continuous-action planning for discounted infinite-horizon nonlinear optimal control with Lipschitz values
- Continuous state dynamic programming via nonexpansive approximation
Cites work
- scientific article; zbMATH DE number 700091 (Why is no real title available?)
- Analysis of a Numerical Dynamic Programming Algorithm Applied to Economic Models
- Approximate policy iteration: a survey and some new methods
- Continuous state dynamic programming via nonexpansive approximation
- Convergence Properties of Policy Iteration
- Dynamic programming and optimal control. Vol. 1.
- Dynamic programming and optimal control. Vol. 2
- Existence of optimal stationary policies in deterministic optimal control
- On the Convergence of Policy Iteration in Stationary Dynamic Programming
- The theory of dynamic programming
- The turnpike property in finite-dimensional nonlinear optimal control
- Turnpike phenomenon and infinite horizon optimal control
- Variable resolution discretization in optimal control
- What you should know about approximate dynamic programming
Cited in
(2)
This page was built for publication: Undiscounted control policy generation for continuous-valued optimal control by approximate dynamic programming
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q5043547)