A spectral analytic comparison of trace-class data augmentation algorithms and their sandwich variants (Q661172)
From MaRDI portal
scientific article
Language | Label | Description | Also known as |
---|---|---|---|
English | A spectral analytic comparison of trace-class data augmentation algorithms and their sandwich variants |
scientific article |
Statements
A spectral analytic comparison of trace-class data augmentation algorithms and their sandwich variants (English)
0 references
21 February 2012
0 references
Given a joint probability density \(f(x,y)\) with respect to measure \(\mu\times \nu\) on the set \(X \times Y\), the data augmentation (DA) algorithm is a Markov chain Monte Carlo technique which is used to explore the marginal probability density \(f_X(x)= \int_Y f(x,y) \nu(dy)\). The Markov chain underlying the algorithm, denoted by \(\{X_n\}_{n=0}^\infty\), is the following. If \(X_n=x\), then \(X_{n+1}\) is simulated in two steps: draw \(Y \sim f_{Y|X}(\cdot|x)\), call the result \(y\), and then draw \(X_{n+1} \sim f_{X|Y}(\cdot|y)\). Although DA is easy to implement it often suffers from slow convergence. The sandwich algorithm is an alternative to DA. Given the same conditions and \(X_n^*=x\), each iteration of the sandwich algorithm is simulated in three steps (one additional step is ``sandwiched'' between two original steps): draw \(Y \sim f_{Y|X}(\cdot|x)\), call the result \(y\), draw \(Y' \sim R(y,\cdot)\), call the result \(y'\), and then draw \(X_{n+1}^* \sim f_{X|Y}(\cdot|y')\). Here, \(R(y,dy')\) is any Markov transition function reversible with respect to \(f_Y(y)\nu(dy)\). If \(R(y,dy')\) is properly chosen, the sandwich algorithm \(\{X_n^*\}_{n=0}^\infty\) can converge much faster than DA while requiring roughly the same computational effort per iteration. Empirical studies pointing to the superiority of the sandwich algorithm abound. The development of confirmatory theoretical results has been slow. It is known that sandwich algorithm always converges at least as fast as the corresponding DA algorithm in the sense that \(\|K^*\| \leq \|K\|\), where \(K\) and \(K^*\) are the Markov operators associated with the DA and sandwich algorithms, respectively, and \(\| \cdot \|\) denotes the operator norm. In this paper, a substantial refinement of this operator norm inequality is developed. Results are applied to a new DA algorithm for Bayesian quantile regression introduced by \textit{H. Kozumi} and \textit{G. Kobayashi} [``Gibbs sampling methods for Bayesian quantile regression'', J. Stat. Comput. Simul. 81, No. 11, 1565--1578 (2011; \url{doi:10.1080/00949655.2010.496117})].
0 references
compact operator
0 references
convergence rate
0 references
eigenvalues
0 references
group action
0 references
Markov chains
0 references
Markov operator
0 references
Monte Carlo method
0 references
operator norm
0 references
positive operator
0 references
0 references
0 references
0 references
0 references