New approximate dynamic programming algorithms for large-scale undiscounted Markov decision processes and their application to optimize a production and distribution system
DOI10.1016/J.EJOR.2015.07.026zbMATH Open1346.90803OpenAlexW2197046949MaRDI QIDQ320866FDOQ320866
Authors: Katsuhisa Ohno, Toshitaka Boh, Koichi Nakade, Takayoshi Tamura
Publication date: 7 October 2016
Published in: European Journal of Operational Research (Search for Journal in Brave)
Full work available at URL: https://doi.org/10.1016/j.ejor.2015.07.026
Recommendations
- The optimal control of just-in-time-based production and distribution systems and performance comparisons with optimized pull systems
- Approximate dynamic programming by practical examples
- Simplex algorithm for countable-state discounted Markov decision processes
- Approximate dynamic programming for stochastic linear control problems on compact state spaces
- Suboptimal policy determination for large-scale Markov decision processes. I: Description and bounds
optimal controlapproximate dynamic programming algorithmscurses of dimensionalityJIT-based production and distribution systemundiscounted Markov decision processes
Large-scale problems in mathematical programming (90C06) Dynamic programming (90C39) Markov and semi-Markov decision processes (90C40)
Cites Work
- Title not available (Why is that?)
- Title not available (Why is that?)
- Title not available (Why is that?)
- Stochastic learning and optimization. A sensitivity-based approach.
- Solving semi-Markov decision problems using average reward reinforcement learning
- Approximate Dynamic Programming
- Approximate policy iteration: a survey and some new methods
- A review of stochastic algorithms with continuous value function approximation and some new approximate policy iteration algorithms for multidimensional continuous applications
- Simulation-based algorithms for Markov decision processes.
- Simulation-based optimization: Parametric optimization techniques and reinforcement learning
- The optimal control of just-in-time-based production and distribution systems and performance comparisons with optimized pull systems
- A comparison of production-line control mechanisms
- Optimal numbers of two kinds of kanbans in a JIT production system
- CONVERGENCE OF SIMULATION-BASED POLICY ITERATION
- Approximate Dynamic Programming via a Smoothed Linear Program
- Computing Optimal Policies for Controlled Tandem Queueing Systems
Cited In (7)
- A performance-centred approach to optimising maintenance of complex systems
- Relevant states and memory in Markov chain bootstrapping and simulation
- Fast heuristic approach for control of complex authentication systems
- A typology and literature review on stochastic multi-echelon inventory models
- Integrated optimization of material supplying, manufacturing, and product distribution: models and fast algorithms
- Sensitivity and covariance in stochastic complementarity problems with an application to north American natural gas markets
- Dynamic assignment of a multi-skilled workforce in job shops: an approximate dynamic programming approach
This page was built for publication: New approximate dynamic programming algorithms for large-scale undiscounted Markov decision processes and their application to optimize a production and distribution system
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q320866)