Multi-armed bandit experiments in the online service economy
From MaRDI portal
Publication:6574679
DOI10.1002/ASMB.2104MaRDI QIDQ6574679FDOQ6574679
Authors: Steven L. Scott
Publication date: 18 July 2024
Published in: Applied Stochastic Models in Business and Industry (Search for Journal in Brave)
Cites Work
- Title not available (Why is that?)
- Asymptotically efficient adaptive allocation rules
- Title not available (Why is that?)
- Title not available (Why is that?)
- Some aspects of the sequential design of experiments
- Finite-time analysis of the multiarmed bandit problem
- Statistical methods for dynamic treatment regimes. Reinforcement learning, causal inference, and personalized medicine
- Thompson sampling: an asymptotically optimal finite-time analysis
- Bandit problems with infinitely many arms
- Learning to optimize via posterior sampling
- Title not available (Why is that?)
- On the likelihood that one unkrown probability exeeds another in view of the evidence of two samples.
Cited In (5)
This page was built for publication: Multi-armed bandit experiments in the online service economy
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6574679)