Asymptotic bias of stochastic gradient search (Q1704136)

From MaRDI portal
Revision as of 04:24, 5 March 2024 by Import240304020342 (talk | contribs) (Set profile property.)
scientific article
Language Label Description Also known as
English
Asymptotic bias of stochastic gradient search
scientific article

    Statements

    Asymptotic bias of stochastic gradient search (English)
    0 references
    0 references
    0 references
    8 March 2018
    0 references
    There is an investigation on the asymptotic behavior of biased stochastic gradient search. The following algorithm is analyzed: \({\theta}_{n+1}\) = \({\theta}_{n}\)- \({\alpha}_{n}({\nabla}f({\theta}_{n})+{\xi}_{n})\) , \(n{\geq}0\). Under a set of assumptions regarding the step-size sequence, the noise and the objective function f , the convergence of the algorithm iterates to a neighborhood of the set of minima, is proved. Upper bounds on the radius of the vicinity are obtained. The results are local, they hold only in case the stated algorithm is stable. The proofs are relaying on the chain-recurrence, Yomdin theorem and Lojasiewicz inequalities. Further , the obtained results are applied to stochastic gradient algorithms with Markovian dynamics and to the asymptotic analysis of a policy-gradient search algorithm for average-cost Markov decision problems. Global versions of the results are presented in the Appendix A and Appendix B of the paper. There is stipulated that an extended version of this article is available at \url{arXiv:1709.00291}.
    0 references
    stochastic gradient search
    0 references
    biased gradient estimation
    0 references
    chain-recurrence
    0 references
    Yomdin theorem
    0 references
    Lojasiewicz inequalities
    0 references

    Identifiers

    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references