Infinite-dimensional gradient-based descent for alpha-divergence minimisation

From MaRDI portal
Publication:2054493

DOI10.1214/20-AOS2035zbMATH Open1486.62065arXiv2005.10618OpenAlexW3204854592MaRDI QIDQ2054493FDOQ2054493


Authors: Kamélia Daudel, Randal Douc, François Portier Edit this on Wikidata


Publication date: 3 December 2021

Published in: The Annals of Statistics (Search for Journal in Brave)

Abstract: This paper introduces the (alpha,Gamma)-descent, an iterative algorithm which operates on measures and performs alpha-divergence minimisation in a Bayesian framework. This gradient-based procedure extends the commonly-used variational approximation by adding a prior on the variational parameters in the form of a measure. We prove that for a rich family of functions Gamma, this algorithm leads at each step to a systematic decrease in the alpha-divergence and derive convergence results. Our framework recovers the Entropic Mirror Descent algorithm and provides an alternative algorithm that we call the Power Descent. Moreover, in its stochastic formulation, the (alpha,Gamma)-descent allows to optimise the mixture weights of any given mixture model without any information on the underlying distribution of the variational parameters. This renders our method compatible with many choices of parameters updates and applicable to a wide range of Machine Learning tasks. We demonstrate empirically on both toy and real-world examples the benefit of using the Power descent and going beyond the Entropic Mirror Descent framework, which fails as the dimension grows.


Full work available at URL: https://arxiv.org/abs/2005.10618




Recommendations




Cites Work


Cited In (3)





This page was built for publication: Infinite-dimensional gradient-based descent for alpha-divergence minimisation

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q2054493)