On the stability of sequential Monte Carlo methods in high dimensions

From MaRDI portal
Publication:2511554




Abstract: We investigate the stability of a Sequential Monte Carlo (SMC) method applied to the problem of sampling from a target distribution on mathbbRd for large d. It is well known that using a single importance sampling step one produces an approximation for the target that deteriorates as the dimension d increases, unless the number of Monte Carlo samples N increases at an exponential rate in d. We show that this degeneracy can be avoided by introducing a sequence of artificial targets, starting from a `simple' density and moving to the one of interest, using an SMC method to sample from the sequence. Using this class of SMC methods with a fixed number of samples, one can produce an approximation for which the effective sample size (ESS) converges to a random variable varepsilonN as dightarrowinfty with 1<varepsilonN<N. The convergence is achieved with a computational cost proportional to Nd2. If varepsilonNllN, we can raise its value by introducing a number of resampling steps, say m (where m is independent of d). In this case, ESS converges to a random variable varepsilonN,m as dightarrowinfty and limmoinftyvarepsilonN,m=N. Also, we show that the Monte Carlo error for estimating a fixed dimensional marginal expectation is of order frac1sqrtN uniformly in d. The results imply that, in high dimensions, SMC algorithms can efficiently control the variability of the importance sampling weights and estimate fixed dimensional marginals at a cost which is less than exponential in d and indicate that, in high dimensions, resampling leads to a reduction in the Monte Carlo error and increase in the ESS.



Cites work


Cited in
(58)






This page was built for publication: On the stability of sequential Monte Carlo methods in high dimensions

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q2511554)