Multi-objective optimization of long-run average and total rewards

From MaRDI portal
Publication:2044201

DOI10.1007/978-3-030-72016-2_13zbMATH Open1467.68094arXiv2010.13566OpenAlexW3147073715MaRDI QIDQ2044201FDOQ2044201


Authors: Tim Quatmann, Joost-Pieter Katoen Edit this on Wikidata


Publication date: 4 August 2021

Abstract: This paper presents an efficient procedure for multi-objective model checking of long-run average reward (aka: mean pay-off) and total reward objectives as well as their combination. We consider this for Markov automata, a compositional model that captures both traditional Markov decision processes (MDPs) as well as a continuous-time variant thereof. The crux of our procedure is a generalization of Forejt et al.'s approach for total rewards on MDPs to arbitrary combinations of long-run and total reward objectives on Markov automata. Experiments with a prototypical implementation on top of the Storm model checker show encouraging results for both model types and indicate a substantial improved performance over existing multi-objective long-run MDP model checking based on linear programming.


Full work available at URL: https://arxiv.org/abs/2010.13566




Recommendations



Cites Work


Cited In (2)





This page was built for publication: Multi-objective optimization of long-run average and total rewards

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q2044201)