Asymptotic bias of stochastic gradient search

From MaRDI portal
Publication:1704136

DOI10.1214/16-AAP1272zbMATH Open1387.49044arXiv1709.00291OpenAlexW2591423585MaRDI QIDQ1704136FDOQ1704136


Authors: Vladislav B. Tadić, Arnaud Doucet Edit this on Wikidata


Publication date: 8 March 2018

Published in: The Annals of Applied Probability (Search for Journal in Brave)

Abstract: The asymptotic behavior of the stochastic gradient algorithm with a biased gradient estimator is analyzed. Relying on arguments based on the dynamic system theory (chain-recurrence) and the differential geometry (Yomdin theorem and Lojasiewicz inequality), tight bounds on the asymptotic bias of the iterates generated by such an algorithm are derived. The obtained results hold under mild conditions and cover a broad class of high-dimensional nonlinear algorithms. Using these results, the asymptotic properties of the policy-gradient (reinforcement) learning and adaptive population Monte Carlo sampling are studied. Relying on the same results, the asymptotic behavior of the recursive maximum split-likelihood estimation in hidden Markov models is analyzed, too.


Full work available at URL: https://arxiv.org/abs/1709.00291




Recommendations




Cites Work


Cited In (17)





This page was built for publication: Asymptotic bias of stochastic gradient search

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q1704136)