Laplace's rule of succession in Information Geometry
From MaRDI portal
Publication:2807529
DOI10.1007/978-3-319-25040-3_34zbMATH Open1396.94034OpenAlexW2962999312MaRDI QIDQ2807529FDOQ2807529
Authors: Y. Ollivier
Publication date: 25 May 2016
Published in: Lecture Notes in Computer Science (Search for Journal in Brave)
Abstract: Laplace's "add-one" rule of succession modifies the observed frequencies in a sequence of heads and tails by adding one to the observed counts. This improves prediction by avoiding zero probabilities and corresponds to a uniform Bayesian prior on the parameter. The canonical Jeffreys prior corresponds to the "add-one-half" rule. We prove that, for exponential families of distributions, such Bayesian predictors can be approximated by taking the average of the maximum likelihood predictor and the emph{sequential normalized maximum likelihood} predictor from information theory. Thus in this case it is possible to approximate Bayesian predictors without the cost of integrating or sampling in parameter space.
Full work available at URL: https://arxiv.org/abs/1503.04304
Recommendations
This page was built for publication: Laplace's rule of succession in Information Geometry
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q2807529)