Adapative importance sampling on discrete Markov chains (Q1305417): Difference between revisions

From MaRDI portal
RedirectionBot (talk | contribs)
Removed claims
RedirectionBot (talk | contribs)
Changed an Item
Property / author
 
Property / author: Keith A. Baggerly / rank
 
Normal rank
Property / reviewed by
 
Property / reviewed by: Jaromír Antoch / rank
 
Normal rank

Revision as of 05:29, 10 February 2024

scientific article
Language Label Description Also known as
English
Adapative importance sampling on discrete Markov chains
scientific article

    Statements

    Adapative importance sampling on discrete Markov chains (English)
    0 references
    0 references
    0 references
    0 references
    0 references
    22 March 2000
    0 references
    In modelling particle transport through the medium, the path of a particle behaves as a transient Markov chain. The authors are interested in characteristics of the particle's movement depending on its starting state, which take the form of a ``score'' accumulated with each transition. The main purpose of this work is to prove that under certain conditions adaptive importance sampling for discrete Markov chains with scoring converges exponentially. Examples presented show that this exponential convergence can occur with a reasonably small number of simulation runs. These assumptions include that the state space is finite, the vector of expected scores conform a linear model and that there are sufficiently many replications of the initial states in the simulation.
    0 references
    adaptive procedures
    0 references
    exponential convergence
    0 references
    Monte Carlo method
    0 references
    particle transport
    0 references
    zero-variance solution
    0 references
    importance sampling
    0 references
    discrete Markov chains
    0 references
    numerical examples
    0 references

    Identifiers

    0 references
    0 references
    0 references
    0 references
    0 references