Dynamic importance sampling for uniformly recurrent Markov chains (Q1774208)
From MaRDI portal
| This is the item page for this Wikibase entity, intended for internal use and editing purposes. Please use this page instead for the normal view: Dynamic importance sampling for uniformly recurrent Markov chains |
scientific article
| Language | Label | Description | Also known as |
|---|---|---|---|
| default for all languages | No label defined |
||
| English | Dynamic importance sampling for uniformly recurrent Markov chains |
scientific article |
Statements
Dynamic importance sampling for uniformly recurrent Markov chains (English)
0 references
29 April 2005
0 references
A dynamic important sampling in the setting of uniformly recurrent Markov chain is suggested for reducing the variance of estimating the rare probability by Monte Carlo, where the reference measure used for rejection principle is time varying. The existence of asymptotically optimal scheme is shown based on a control theory approach. Numerical examples are illustrated in contrast to the standard scheme.
0 references
asymptotic optimality
0 references
important sampling
0 references
Monte Carlo simulation
0 references
rare event
0 references
weak convergence
0 references
0 references
0 references
0 references
0 references
0 references
0.8421599268913269
0 references
0.8117820620536804
0 references
0.8004481196403503
0 references
0.8001993894577026
0 references
0.7988709211349487
0 references