Nonuniqueness versus uniqueness of optimal policies in convex discounted Markov decision processes (Q2375462): Difference between revisions

From MaRDI portal
Importer (talk | contribs)
Created a new Item
 
ReferenceBot (talk | contribs)
Changed an Item
 
(6 intermediate revisions by 5 users not shown)
Property / author
 
Property / author: Raúl Montes-De-oca / rank
Normal rank
 
Property / author
 
Property / author: Raúl Montes-De-oca / rank
 
Normal rank
Property / Wikidata QID
 
Property / Wikidata QID: Q59002487 / rank
 
Normal rank
Property / MaRDI profile type
 
Property / MaRDI profile type: MaRDI publication profile / rank
 
Normal rank
Property / full work available at URL
 
Property / full work available at URL: https://doi.org/10.1155/2013/271279 / rank
 
Normal rank
Property / OpenAlex ID
 
Property / OpenAlex ID: W2001225281 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Conditions for the uniqueness of optimal policies of discounted Markov decision processes / rank
 
Normal rank
Property / cites work
 
Property / cites work: Average Optimality in Markov Control Processes via Discounted-Cost Problems and Linear Programming / rank
 
Normal rank
Property / cites work
 
Property / cites work: Note—A Note on Dynamic Programming with Unbounded Rewards / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q3994363 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q4844751 / rank
 
Normal rank
links / mardi / namelinks / mardi / name
 

Latest revision as of 13:28, 6 July 2024

scientific article
Language Label Description Also known as
English
Nonuniqueness versus uniqueness of optimal policies in convex discounted Markov decision processes
scientific article

    Statements

    Nonuniqueness versus uniqueness of optimal policies in convex discounted Markov decision processes (English)
    0 references
    14 June 2013
    0 references
    Summary: From the classical point of view, it is important to determine if in a Markov decision process (MDP), besides their existence, the uniqueness of the optimal policies is guaranteed. It is well known that uniqueness does not always hold in optimization problems (for instance, in linear programming). On the other hand, in such problems it is possible for a slight perturbation of the functional cost to restore the uniqueness. In this paper, it is proved that the value functions of an MDP and its cost perturbed version stay close, under adequate conditions, which in some sense is a priority. We are interested in the stability of Markov decision processes with respect to the perturbations of the cost-as-you-go function.
    0 references

    Identifiers