Complete stability analysis of a heuristic approximate dynamic programming control design
From MaRDI portal
Abstract: This paper provides new stability results for Action-Dependent Heuristic Dynamic Programming (ADHDP), using a control algorithm that iteratively improves an internal model of the external world in the autonomous system based on its continuous interaction with the environment. We extend previous results by ADHDP control to the case of general multi-layer neural networks with deep learning across all layers. In particular, we show that the introduced control approach is uniformly ultimately bounded (UUB) under specific conditions on the learning rates, without explicit constraints on the temporal discount factor. We demonstrate the benefit of our results to the control of linear and nonlinear systems, including the cart-pole balancing problem. Our results show significantly improved learning and control performance as compared to the state-of-art.
Recommendations
- A boundedness result for the direct heuristic dynamic programming
- Robust reinforcement learning control with static and dynamic stability
- Adaptive dynamic programming for control. Algorithms and stability
- Performance evaluation of direct heuristic dynamic programming using control-theoretic measures
- Convergence analysis of the deep neural networks based globalized dual heuristic programming
Cites work
- scientific article; zbMATH DE number 1321699 (Why is no real title available?)
- A boundedness result for the direct heuristic dynamic programming
- A nearly optimal control scheme using adaptive dynamic programming based on the generalized fuzzy hyperbolic model
- Adaptive dynamic programming for control. Algorithms and stability
- An iterative adaptive dynamic programming method for solving a class of nonlinear zero-sum differential games
- Approximate dynamic programming. Solving the curses of dimensionality
- Model-free \(Q\)-learning designs for linear discrete-time zero-sum games with application to \(H^\infty\) control
- Nearly optimal control laws for nonlinear systems with saturating actuators using a neural network HJB approach
- Neural network control of nonlinear discrete-time systems.
- Optimal control of unknown nonaffine nonlinear discrete-time systems based on adaptive dynamic programming
- Stability of dynamical systems. Continuous, discontinuous, and discrete systems
- Universal approximation bounds for superpositions of a sigmoidal function
Cited in
(26)- Heuristic dynamic programming-based learning control for discrete-time disturbed multi-agent systems
- Robust reinforcement learning control with static and dynamic stability
- Application of heuristic programming to dynamic system stabilization
- Convergence analysis of the deep neural networks based globalized dual heuristic programming
- Q-learning solution for optimal consensus control of discrete-time multiagent systems using reinforcement learning
- A stabilizing reinforcement learning approach for sampled systems with partially unknown models
- Attitude control with auxiliary structure based on adaptive dynamic programming for reentry vehicles
- Performance evaluation of direct heuristic dynamic programming using control-theoretic measures
- Model‐free incremental adaptive dynamic programming based approximate robust optimal regulation
- Dynamic event-triggered control for discrete-time nonlinear Markov jump systems using policy iteration-based adaptive dynamic programming
- Model‐free optimal tracking over finite horizon using adaptive dynamic programming
- Neural‐network‐based control for discrete‐time nonlinear systems with denial‐of‐service attack: The adaptive event‐triggered case
- A Q-learning predictive control scheme with guaranteed stability
- Event-triggered adaptive dynamic programming for discrete-time multi-player games
- Neural-network-based discounted optimal control via an integrated value iteration with accuracy guarantee
- Data-driven approximate Q-learning stabilization with optimality error bound analysis
- Prioritized experience replay based reinforcement learning for adaptive tracking control of autonomous underwater vehicle
- Data-driven approximate value iteration with optimality error bound analysis
- Three bounded proofs for nonlinear multi-input multi-output approximate dynamic programming based on the Lyapunov stability theory
- Toward reliable designs of data-driven reinforcement learning tracking control for Euler-Lagrange systems
- Data-driven adaptive dynamic programming for partially observable nonzero-sum games via \(Q\)-learning method
- Optimal consensus control for double‐integrator multiagent systems with unknown dynamics using adaptive dynamic programming
- Optimal reconstruction of constrained Janbu method with ADP and non-integral safety factor
- Adaptive event-triggered control and observer design for discrete-time nonlinear Markov jump systems with DoS attacks using policy iteration-based adaptive dynamic programming
- Spacecraft output feedback attitude control based on extended state observer and adaptive dynamic programming
- A boundedness result for the direct heuristic dynamic programming
This page was built for publication: Complete stability analysis of a heuristic approximate dynamic programming control design
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q894322)