Gaussian approximations for probability measures on R^d

From MaRDI portal
Publication:4636355

DOI10.1137/16M1105384zbMATH Open1390.60022arXiv1611.08642MaRDI QIDQ4636355FDOQ4636355


Authors: Yulong Lu, Hendrik Weber, A. M. Stuart Edit this on Wikidata


Publication date: 19 April 2018

Published in: SIAM/ASA Journal on Uncertainty Quantification (Search for Journal in Brave)

Abstract: This paper concerns the approximation of probability measures on mathbfRd with respect to the Kullback-Leibler divergence. Given an admissible target measure, we show the existence of the best approximation, with respect to this divergence, from certain sets of Gaussian measures and Gaussian mixtures. The asymptotic behavior of such best approximations is then studied in the small parameter limit where the measure concentrates; this asymptotic behaviour is characterized using Gamma-convergence. The theory developed is then applied to understanding the frequentist consistency of Bayesian inverse problems. For a fixed realization of noise, we show the asymptotic normality of the posterior measure in the small noise limit. Taking into account the randomness of the noise, we prove a Bernstein-Von Mises type result for the posterior measure.


Full work available at URL: https://arxiv.org/abs/1611.08642




Recommendations




Cites Work


Cited In (22)

Uses Software





This page was built for publication: Gaussian approximations for probability measures on \(\mathbb R^d\)

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q4636355)