A decentralized partially observable Markov decision model with action duration for goal recognition in real time strategy games (Q2403912): Difference between revisions

From MaRDI portal
Import240304020342 (talk | contribs)
Set profile property.
Set OpenAlex properties.
Property / full work available at URL
 
Property / full work available at URL: https://doi.org/10.1155/2017/4580206 / rank
 
Normal rank
Property / OpenAlex ID
 
Property / OpenAlex ID: W2735072884 / rank
 
Normal rank

Revision as of 03:01, 20 March 2024

scientific article
Language Label Description Also known as
English
A decentralized partially observable Markov decision model with action duration for goal recognition in real time strategy games
scientific article

    Statements

    A decentralized partially observable Markov decision model with action duration for goal recognition in real time strategy games (English)
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    12 September 2017
    0 references
    Summary: Multiagent goal recognition is a tough yet important problem in many real time strategy games or simulation systems. Traditional modeling methods either are in great demand of detailed agents' domain knowledge and training dataset for policy estimation or lack clear definition of action duration. To solve the above problems, we propose a novel Dec-POMDM-T model, combining the classic Dec-POMDP, an observation model for recognizer, joint goal with its termination indicator, and time duration variables for actions with action termination variables. In this paper, a model-free algorithm named cooperative colearning based on Sarsa is used. Considering that Dec-POMDM-T usually encounters multiagent goal recognition problems with different sorts of noises, partially missing data, and unknown action durations, the paper exploits the SIS PF with resampling for inference under the dynamic Bayesian network structure of Dec-POMDM-T. In experiments, a modified predator-prey scenario is adopted to study multiagent joint goal recognition problem, which is the recognition of the joint target shared among cooperative predators. Experiment results show that (a) Dec-POMDM-T works effectively in multiagent goal recognition and adapts well to dynamic changing goals within agent group; (b) Dec-POMDM-T outperforms traditional Dec-MDP-based methods in terms of precision, recall, and \(F\)-measure.
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references