A semi-Markov decision model for recognizing the destination of a maneuvering agent in real time strategy games (Q1792771): Difference between revisions
From MaRDI portal
Set profile property. |
ReferenceBot (talk | contribs) Changed an Item |
||
(One intermediate revision by one other user not shown) | |||
Property / full work available at URL | |||
Property / full work available at URL: https://doi.org/10.1155/2016/1907971 / rank | |||
Normal rank | |||
Property / OpenAlex ID | |||
Property / OpenAlex ID: W2236759984 / rank | |||
Normal rank | |||
Property / cites work | |||
Property / cites work: Q4411146 / rank | |||
Normal rank | |||
Property / cites work | |||
Property / cites work: Learning hierarchical task network domains from partially observed plan traces / rank | |||
Normal rank | |||
Property / cites work | |||
Property / cites work: Q3655273 / rank | |||
Normal rank | |||
Property / cites work | |||
Property / cites work: Location-Based Reasoning about Complex Multi-Agent Behavior / rank | |||
Normal rank |
Latest revision as of 21:26, 16 July 2024
scientific article
Language | Label | Description | Also known as |
---|---|---|---|
English | A semi-Markov decision model for recognizing the destination of a maneuvering agent in real time strategy games |
scientific article |
Statements
A semi-Markov decision model for recognizing the destination of a maneuvering agent in real time strategy games (English)
0 references
12 October 2018
0 references
Summary: Recognizing destinations of a maneuvering agent is important in real time strategy games. Because finding path in an uncertain environment is essentially a sequential decision problem, we can model the maneuvering process by the Markov decision process (MDP). However, the MDP does not define an action duration. In this paper, we propose a novel semi-Markov decision model (SMDM). In the SMDM, the destination is regarded as a hidden state, which affects selection of an action; the action is affiliated with a duration variable, which indicates whether the action is completed. We also exploit a Rao-Blackwellised particle filter (RBPF) for inference under the dynamic Bayesian network structure of the SMDM. In experiments, we simulate agents' maneuvering in a combat field and employ agents' traces to evaluate the performance of our method. The results show that the SMDM outperforms another extension of the MDP in terms of precision, recall, and \(F\)-measure. Destinations are recognized efficiently by our method no matter whether they are changed or not. Additionally, the RBPF infer destinations with smaller variance and less time than the SPF. The average failure rates of the RBPF are lower when the number of particles is not enough.
0 references