How Fast Is the Bandit?
From MaRDI portal
Publication:3506304
DOI10.1080/07362990802007202zbMath1416.62456arXivmath/0510351OpenAlexW2056330474MaRDI QIDQ3506304
Damien Lamberton, Gilles Pagès
Publication date: 12 June 2008
Published in: Stochastic Analysis and Applications (Search for Journal in Brave)
Full work available at URL: https://arxiv.org/abs/math/0510351
Learning and adaptive systems in artificial intelligence (68T05) Adaptive control/observation systems (93C40) Stochastic approximation (62L20)
Related Items (4)
Regret bounds for Narendra-Shapiro bandit algorithms ⋮ On ergodic two-armed bandits ⋮ Convergence in models with bounded expected relative hazard rates ⋮ Nonlinear randomized urn models: a stochastic approximation viewpoint
Cites Work
- Unnamed Item
- Nonconvergence to unstable points in urn models and stochastic approximations
- Weak convergence rates for stochastic approximation with application to multiple targets and simulated annealing
- Asymptotics in randomized urn models
- When can the two-armed bandit algorithm be trusted?
- On the linear model with two absorbing barriers
- Learning Automata - A Survey
- Decreasing step Stochastic algorithms: a.s. behaviour of weighted empirical measures
- Use of Stochastic Automata for Parameter Self-Optimization with Multimodal Performance Criteria
This page was built for publication: How Fast Is the Bandit?