Average optimality for continuous-time Markov decision processes with a policy iteration approach (Q2465179): Difference between revisions

From MaRDI portal
Import240304020342 (talk | contribs)
Set profile property.
ReferenceBot (talk | contribs)
Changed an Item
 
(One intermediate revision by one other user not shown)
Property / full work available at URL
 
Property / full work available at URL: https://doi.org/10.1016/j.jmaa.2007.06.071 / rank
 
Normal rank
Property / OpenAlex ID
 
Property / OpenAlex ID: W2073471284 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Equivalence of exponential ergodicity and \(L^ 2\)-exponential convergence for Markov chains. / rank
 
Normal rank
Property / cites work
 
Property / cites work: Continuous time control of Markov processes on an arbitrary state space: Discounted rewards / rank
 
Normal rank
Property / cites work
 
Property / cites work: Exponential and uniform ergodicity of Markov processes / rank
 
Normal rank
Property / cites work
 
Property / cites work: Drift and monotonicity conditions for continuous-time controlled markov chains with an average criterion / rank
 
Normal rank
Property / cites work
 
Property / cites work: A note on optimality conditions for continuous-time Markov decision processes with average cost criterion / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q5463021 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q4534402 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Average optimality for continuous-time Markov decision processes in Polish spaces / rank
 
Normal rank
Property / cites work
 
Property / cites work: Bias Optimality in Controlled Queueing Systems / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q4891056 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q4255598 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q3266141 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Nondiscounted Continuous Time Markovian Decision Process with Countable State Space / rank
 
Normal rank
Property / cites work
 
Property / cites work: On maximal rewards and \(\varepsilon\)-optimal policies in continuous time Markov decision chains / rank
 
Normal rank
Property / cites work
 
Property / cites work: A note on bias optimality in controlled queueing systems / rank
 
Normal rank
Property / cites work
 
Property / cites work: Computable exponential convergence rates for stochastically ordered Markov processes / rank
 
Normal rank
Property / cites work
 
Property / cites work: Finite state continuous time Markov decision processes with an infinite planning horizon / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q4223191 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Criteria for ergodicity, exponential ergodicity and strong ergodicity of Markov processes / rank
 
Normal rank
Property / cites work
 
Property / cites work: On Homogeneous Markov Models with Continuous Time and Finite or Countable State Space / rank
 
Normal rank
Property / cites work
 
Property / cites work: Continuous time Markov decision programming with average reward criterion and unbounded reward rate / rank
 
Normal rank
Property / cites work
 
Property / cites work: Bias optimality and strong \(n\) \((n= -1,0)\) discount optimality for Markov decision processes / rank
 
Normal rank
Property / cites work
 
Property / cites work: Average optimality inequality for continuous-time Markov decision processes in Polish spaces / rank
 
Normal rank
Property / cites work
 
Property / cites work: Markov Decision Processes with Variance Minimization: A New Condition and Approach / rank
 
Normal rank
Property / cites work
 
Property / cites work: Unbounded cost Markov decision processes with limsup and liminf average criteria: new conditions / rank
 
Normal rank
Property / cites work
 
Property / cites work: Another Set of Conditions for Strong<i>n</i>(<i>n</i> = −1, 0) Discount Optimality in Markov Decision Processes / rank
 
Normal rank
Property / cites work
 
Property / cites work: Another set of conditions for Markov decision processes with average sample-path costs / rank
 
Normal rank

Latest revision as of 13:51, 27 June 2024

scientific article
Language Label Description Also known as
English
Average optimality for continuous-time Markov decision processes with a policy iteration approach
scientific article

    Statements

    Average optimality for continuous-time Markov decision processes with a policy iteration approach (English)
    0 references
    0 references
    8 January 2008
    0 references
    The paper deals with the average expected reward criterion for continuous-time Markov decision processes in general state and action spaces. The transition rates of underlying continuous-time jump Markov processes are allowed to be unbounded, and the reward rates may have neither upper nor lower bounds. The author gives conditions on the system's primitive data and under which he proves the existence of the average reward optimality equation and an average optimal stationary policy. Also, under proposed conditions the author ensures the existence of \(\epsilon\)-average optimal stationary polices. Moreover, the author studies some properties of average optimal stationary polices. The author not only establishes another average optimality equation on an average optimal stationary policy, but also presents an interesting ``martingale characterization'' of such a policy. The approach presented in the paper is based on the policy iteration algorithm and is different from those (``vanishing discounting factor approach'', ``optimality inequality approach'') usually used in the literature. References contain \(31\) items.
    0 references
    continuous-time Markov decision processes
    0 references
    policy iteration algorithm
    0 references
    average criterion
    0 references
    optimality equation
    0 references
    optimal stationary policy
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references

    Identifiers