Stochastic control via direct comparison
DOI10.1007/s10626-010-0093-4zbMath1210.93081MaRDI QIDQ633815
De-Xin Wang, Yi-Fan Xu, Tao Lü, Cao, Xiren
Publication date: 30 March 2011
Published in: Discrete Event Dynamic Systems (Search for Journal in Brave)
Full work available at URL: https://doi.org/10.1007/s10626-010-0093-4
dynamic programming; Markov decision processes; Poisson equation; perturbation analysis; HJB equation; performance potentials; sensitivity-based optimization
60J25: Continuous-time Markov processes on general state spaces
49L20: Dynamic programming in optimal control and differential games
90C15: Stochastic programming
93C70: Time-scale analysis and singular perturbations in control/observation systems
90C39: Dynamic programming
93E20: Optimal stochastic control
Related Items
Uses Software
Cites Work
- Stochastic theory of inventory control
- Policy iteration for customer-average performance optimization of closed queueing systems
- From perturbation analysis to Markov decision processes and reinforcement learning
- A Liapounov bound for solutions of the Poisson equation
- A single sample path-based performance sensitivity formula for Markov chains
- Stability of Markovian processes III: Foster–Lyapunov criteria for continuous-time processes
- Improved Dynamic Programming Methods for Optimal Control of Lumped-Parameter Stochastic Systems
- The Linear Programming Approach to Approximate Dynamic Programming
- The policy iteration algorithm for average reward Markov decision processes with general state space
- The $n$th-Order Bias Optimality for Multichain Markov Decision Processes
- The Potential Structure of Sample Paths and Performance Sensitivities of Markov Systems
- Relationship Between Perturbation Realization Factors With Queueing Models and Markov Models
- Approximate Dynamic Programming
- A Numerical Method for Solving Singular Stochastic Control Problems
- Discrete Dynamic Programming with Sensitive Discount Optimality Criteria
- Perturbation theory and finite Markov chains
- Handbook of Markov decision processes. Methods and applications
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item