Reading policies for joins: an asymptotic analysis
From MaRDI portal
Publication:2467117
DOI10.1214/105051606000000646zbMATH Open1163.90789arXivmath/0703019OpenAlexW3101031228MaRDI QIDQ2467117FDOQ2467117
Nariankadu D. Shyamalkumar, Ralph P. Russo
Publication date: 18 January 2008
Published in: The Annals of Applied Probability (Search for Journal in Brave)
Abstract: Suppose that observations are made from the distribution and from the distribution . Associate with each pair, from and from , a nonnegative score . An optimal reading policy is one that yields a sequence that maximizes , the expected sum of the observed scores, uniformly in . The alternating policy, which switches between the two sources, is the optimal nonadaptive policy. In contrast, the greedy policy, which chooses its source to maximize the expected gain on the next step, is shown to be the optimal policy. Asymptotics are provided for the case where the and distributions are discrete and according as or not (i.e., the observations match). Specifically, an invariance result is proved which guarantees that for a wide class of policies, including the alternating and the greedy, the variable M(n) obeys the same CLT and LIL. A more delicate analysis of the sequence and the sample paths of M(n), for both alternating and greedy, reveals the slender sense in which the latter policy is asymptotically superior to the former, as well as a sense of equivalence of the two and robustness of the former.
Full work available at URL: https://arxiv.org/abs/math/0703019
Central limit and other weak theorems (60F05) Strong limit theorems (60F15) Stopping times; optimal stopping problems; gambling theory (60G40) Markov and semi-Markov decision processes (90C40)
Cites Work
- Approximation Theorems of Mathematical Statistics
- Title not available (Why is that?)
- Title not available (Why is that?)
- Probability
- Title not available (Why is that?)
- On the Gittins index for multiarmed bandits
- Extensions of the multiarmed bandit problem: The discounted case
- Tauberian Theory
- Title not available (Why is that?)
- Title not available (Why is that?)
- Title not available (Why is that?)
- Some results on increments of the Wiener process with applications to lag sums of i.i.d. random variables
- Sample path optimality for a Markov optimization problem
- Dynamic productivity improvement in a model with multiple processes
- Complete convergence of triangular arrays and the law of the iterated logarithm for U-statistics
- Upper and lower functions for martingales and mixing processes
- On the convergence of the empirical mass function
- Title not available (Why is that?)
- Optimal policies to obtain the most join results
Recommendations
- Optimal policies to obtain the most join results π π
- Optimal adaptive policies for sequential allocation problems π π
- Title not available (Why is that?) π π
- Asymptotically efficient adaptive allocation rules π π
- Optimal Online Selection of an Alternating Subsequence: A Central Limit Theorem π π
This page was built for publication: Reading policies for joins: an asymptotic analysis
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q2467117)