Block-successive approximation for a discounted Markov decision model (Q2265958): Difference between revisions
From MaRDI portal
Set OpenAlex properties. |
ReferenceBot (talk | contribs) Changed an Item |
||
Property / cites work | |||
Property / cites work: Contraction Mappings in the Theory Underlying Dynamic Programming / rank | |||
Normal rank | |||
Property / cites work | |||
Property / cites work: Q4427313 / rank | |||
Normal rank | |||
Property / cites work | |||
Property / cites work: Letter to the Editor—A Test for Suboptimal Actions in Markovian Decision Problems / rank | |||
Normal rank | |||
Property / cites work | |||
Property / cites work: Some Bounds for Discounted Sequential Decision Processes / rank | |||
Normal rank | |||
Property / cites work | |||
Property / cites work: Q4173220 / rank | |||
Normal rank | |||
Property / cites work | |||
Property / cites work: Q5342712 / rank | |||
Normal rank |
Latest revision as of 17:12, 14 June 2024
scientific article
Language | Label | Description | Also known as |
---|---|---|---|
English | Block-successive approximation for a discounted Markov decision model |
scientific article |
Statements
Block-successive approximation for a discounted Markov decision model (English)
0 references
1985
0 references
We suggest a new successive approximation method to compute the optimal discounted reward for finite state and action, discrete time, discounted Markov decision chains. The method is based on a block partitioning of the (stochastic) matrices corresponding to the stationary policies. The method is particularly attractive when the transition matrices are jointly nearly decomposable or nearly completely decomposable.
0 references
successive approximation
0 references
optimal discounted reward
0 references
finite state and action, discrete time, discounted Markov decision chains
0 references
block partitioning
0 references
stationary policies
0 references