Discount-isotone policies for Markov decision processes (Q1106106)
From MaRDI portal
scientific article
Language | Label | Description | Also known as |
---|---|---|---|
English | Discount-isotone policies for Markov decision processes |
scientific article |
Statements
Discount-isotone policies for Markov decision processes (English)
0 references
1988
0 references
Consider a discounted Markov decision process with state space \(I\subseteq {\mathbb{Z}}^ m,\) finite action sets \(K_ i\), transition probabilities \(p=p(i,j,k)\) and non-negative, bounded rewards \(r=r(i,k)\), where \(i,j\in I\), \(k\in K_ i\). Let \(K=\cup_{i\in I}K_ i\). Suppose further that I(K) is endowed with a partial order \(\succsim^*\) \((\succsim)\). Let \(Q=\{\delta_{\rho}\), \(\rho\in [0,1)\}\) be a collection of optimal, stationary policies: one for each discount factor \(\rho\). Q is called D-isotone if \(\tau\geq \rho \Rightarrow \delta_{\tau}(i)\succsim \delta_{\rho}(i)\) for all \(i\in I\), \(\tau,\rho\in [0,1)\). Some sufficient conditions are given (in the form of generalized isotonicity and submodularity requirements on r and p) for the existence of a D-isotone set in the case when \((K,\succsim)\) is a lattice. Structural and computational applications are then illustrated.
0 references
isotone policies
0 references
inventory control
0 references
discounted Markov decision process
0 references
finite action sets
0 references
submodularity
0 references
existence of a D-isotone set
0 references