The multi-armed bandit problem with covariates

From MaRDI portal
Publication:355096

DOI10.1214/13-AOS1101zbMATH Open1360.62436arXiv1110.6084OpenAlexW3100895096WikidataQ56675681 ScholiaQ56675681MaRDI QIDQ355096FDOQ355096


Authors: Vianney Perchet, Philippe Rigollet Edit this on Wikidata


Publication date: 24 July 2013

Published in: The Annals of Statistics (Search for Journal in Brave)

Abstract: We consider a multi-armed bandit problem in a setting where each arm produces a noisy reward realization which depends on an observable random covariate. As opposed to the traditional static multi-armed bandit problem, this setting allows for dynamically changing rewards that better describe applications where side information is available. We adopt a nonparametric model where the expected rewards are smooth functions of the covariate and where the hardness of the problem is captured by a margin parameter. To maximize the expected cumulative reward, we introduce a policy called Adaptively Binned Successive Elimination (abse) that adaptively decomposes the global problem into suitably "localized" static bandit problems. This policy constructs an adaptive partition using a variant of the Successive Elimination (se) policy. Our results include sharper regret bounds for the se policy in a static bandit problem and minimax optimal regret bounds for the abse policy in the dynamic problem.


Full work available at URL: https://arxiv.org/abs/1110.6084




Recommendations




Cites Work


Cited In (42)





This page was built for publication: The multi-armed bandit problem with covariates

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q355096)