Some aspects of the sequential design of experiments

From MaRDI portal
Publication:5817009

DOI10.1090/S0002-9904-1952-09620-8zbMath0049.37009OpenAlexW1998498767WikidataQ29040948 ScholiaQ29040948MaRDI QIDQ5817009

Herbert Robbins

Publication date: 1952

Published in: Bulletin of the American Mathematical Society (Search for Journal in Brave)

Full work available at URL: https://doi.org/10.1090/s0002-9904-1952-09620-8



Related Items

Functional Sequential Treatment Allocation, A Note on the Minimax Solution for the Two-Stage Group Testing Problem, A linear response bandit problem, Smoothness-Adaptive Contextual Bandits, Optimistic Gittins Indices, Incentivizing Exploration with Heterogeneous Value of Money, EXPLORATION–EXPLOITATION POLICIES WITH ALMOST SURE, ARBITRARILY SLOW GROWING ASYMPTOTIC REGRET, MULTI-ARMED BANDITS WITH COVARIATES:THEORY AND APPLICATIONS, Unnamed Item, Ranking and Selection with Covariates for Personalized Decision Making, Sequentielle Versuchspläne, Regret bounds for Narendra-Shapiro bandit algorithms, Unnamed Item, Dynamic Inventory Control with Fixed Setup Costs and Unknown Discrete Demand Distribution, Continuous Assortment Optimization with Logit Choice Probabilities and Incomplete Information, Dynamic Learning and Decision Making via Basis Weight Vectors, Improved algorithms for bandit with graph feedback via regret decomposition, Topp-Leone distribution with an application to binomial sampling, Encounters with Martingales in Statistics and Stochastic Optimization, Asymptotic optimality of myopic ranking and selection procedures, Unnamed Item, Unnamed Item, Unnamed Item, How signaling conventions are established, Game-theoretic statistics and safe anytime-valid inference, Recent advances in reinforcement learning in finance, Treatment recommendation with distributional targets, Response-adaptive randomization in clinical trials: from myths to practical considerations, Empirical Gittins index strategies with \(\varepsilon\)-explorations for multi-armed bandit problems, Consumer strategy, vendor strategy and equilibrium in duopoly markets with production costs, Relaxing the i.i.d. assumption: adaptively minimax optimal regret via root-entropic regularization, Optimal Exploration–Exploitation in a Multi-armed Bandit Problem with Non-stationary Rewards, Transfer learning for contextual multi-armed bandits, Multi-armed linear bandits with latent biases, GROUP SEQUENTIAL TESTS WITH OUTCOME-DEPENDENT TREATMENT ASSIGNMENT, Learning the distribution with largest mean: two bandit frameworks, Tuning Bandit Algorithms in Stochastic Environments, Following the Perturbed Leader to Gamble at Multi-armed Bandits, MNL-Bandit: A Dynamic Learning Approach to Assortment Selection, An analytical approximation and a neural network model for optimal sample size in vendor selection, Dynamic Inventory and Price Controls Involving Unknown Demand on Discrete Nonperishable Items, Learning in Combinatorial Optimization: What and How to Explore, Unnamed Item, Efficient Adaptive Randomization and Stopping Rules in Multi-arm Clinical Trials for Testing a New Treatment, Stochastic approximation, Unnamed Item, Adaptive designs for binary treatment responses in phase III clinical trials: controversies and progress, An overview of sequential methods and their application in clinical trials, Unnamed Item, Unnamed Item, Adaptive Incentive-Compatible Sponsored Search Auction, Sequential Shortest Path Interdiction with Incomplete Information, An asymptotically optimal heuristic for general nonstationary finite-horizon restless multi-armed, multi-action bandits, An introduction to designh optimality with an overview of the literature, Randomized allocation with arm elimination in a bandit problem with covariates, Doubly adaptive biased coin designs with delayed responses, Two-Armed Restless Bandits with Imperfect Information: Stochastic Control and Indexability, On Incomplete Learning and Certainty-Equivalence Control, Contribution of Milton Sobel in Selection Problem Following Ethical Allocation, Derivative-free optimization methods, Pure Exploration in Multi-armed Bandits Problems, Nonasymptotic sequential tests for overlapping hypotheses applied to near-optimal arm identification in bandit models, The theory of dynamic programming, Statistical Inference for Online Decision Making via Stochastic Gradient Descent, Dynamic allocation policies for the finite horizon one armed bandit problem, Benchmark and Survey of Automated Machine Learning Frameworks, A class of adaptive designs, Sequential design with applications to the trim-loss problem, Adaptive Policies for Sequential Sampling under Incomplete Information and a Cost Constraint, On the Bias, Risk, and Consistency of Sample Means in Multi-armed Bandits, Two-Armed Bandit Strategies that Discount Past and Future, An adaptive allocation for continuous response using Wilcoxon-Mann-Whitney score, A non-parametric solution to the multi-armed bandit problem with covariates, Batched bandit problems, The time until the final zero crossing of random sums with application to nonparametric bandit theory, Online machine learning algorithms to optimize performances of complex wireless communication systems, Asymptotic properties of doubly adaptive biased coin designs for multitreatment clinical trials., Woodroofe's one-armed bandit problem revisited, A new family of covariate-adjusted response adaptive designs and their properties, Bandit and covariate processes, with finite or non-denumerable set of arms, Response-adaptive designs for clinical trials: simultaneous learning from multiple patients, Asymptotic properties of covariate-adjusted response-adaptive designs, Some problems of optimal sampling strategy, D-Wave and predecessors: From simulated to quantum annealing, Lipschitzness is all you need to tame off-policy generative adversarial imitation learning, The multi-armed bandit problem with covariates, Learning dynamic algorithm portfolios, Kullback-Leibler upper confidence bounds for optimal sequential allocation, An index-based deterministic convergent optimal algorithm for constrained multi-armed bandit problems, On learning and branching: a survey, Limits for partial maxima of Gaussian random vectors, Robustness of stochastic bandit policies, Optimal experimental design for a class of bandit problems, Combining multiple strategies for multiarmed bandit problems and asymptotic optimality, Gaussian two-armed bandit and optimization of batch data processing, Bandit-based Monte-Carlo structure learning of probabilistic logic programs, One optional observation inflates \(\alpha\) by \({100/\sqrt{n}}\) per cent, Algorithm portfolio selection as a bandit problem with unbounded losses, A perpetual search for talents across overlapping generations: a learning process, An asymptotically optimal policy for finite support models in the multiarmed bandit problem, Truthful learning mechanisms for multi-slot sponsored search auctions with externalities, Pure exploration in finitely-armed and continuous-armed bandits, Controlling unknown linear dynamics with bounded multiplicative regret, The \(K\)-armed dueling bandits problem, Strategic learning in teams, Optimal Bayesian strategies for the infinite-armed Bernoulli bandit, A dynamic programming strategy to balance exploration and exploitation in the bandit problem, Regret bounds for sleeping experts and bandits, A common value experimentation with multiarmed bandits, Herbert Robbins and sequential analysis, Optimal strategies for a class of sequential control problems with precedence relations, Randomized prediction of individual sequences, Finding minimax strategy and minimax risk in a random environment (the two-armed bandit problem), Generative adversarial networks are special cases of artificial curiosity (1990) and also closely related to predictability minimization (1991), Multiclass classification, information, divergence and surrogate risk, Learning the Krepsian state: exploration through consumption, A hybrid breakout local search and reinforcement learning approach to the vertex separator problem, MSO: a framework for bound-constrained black-box global optimization algorithms, Convergence in human decision-making dynamics, BIAS CALCULATIONS FOR ADAPTIVE URN DESIGNS, Machine learning for optimal blackjack counting strategies, Exploration and correlation, Optimal adaptive generalized Pólya urn design for multi-arm clinical trials, Arbitrary side observations in bandit problems, Randomized allocation with nonparametric estimation for contextual multi-armed bandits with delayed rewards, A note on infinite-armed Bernoulli bandit problems with generalized beta prior distributions, Efficient crowdsourcing of unknown experts using bounded multi-armed bandits, A general latent assignment approach for modeling psychological contaminants, Asymptotic theorems of sequential estimation-adjusted urn models, SWITCHING MECHANISMS IN A GENERALIZED INFORMATION SYSTEM, The apparent conflict between estimation and control - a survey of the two-armed bandit problem, Sequential monitoring of response-adaptive randomized clinical trials, Boundary crossing probabilities for general exponential families, An online algorithm for the risk-aware restless bandit, Implementing optimal allocation in clinical trials with multiple endpoints, Psychometric engineering as art, Stochastic approximation: from statistical origin to big-data, multidisciplinary applications, Playing with data -- or how to discourage questionable research practices and stimulate researchers to do things right, Concentration bounds for empirical conditional value-at-risk: the unbounded case, Mechanisms with learning for stochastic multi-armed bandit problems, Signaling Games, Certainty equivalence control with forcing: Revisited, Comparison of two Bernoulli processes by multiple stage sampling using Bayesian decision theory, A bad arm existence checking problem: how to utilize asymmetric problem structure?, Exploration-exploitation tradeoff using variance estimates in multi-armed bandits, Gorthaur-EXP3: bandit-based selection from a portfolio of recommendation algorithms balancing the accuracy-diversity dilemma, Active Online Learning in the Binary Perceptron Problem, Multi-armed bandit with sub-exponential rewards, New classes of stochastic control processes, The role of forgetting in the evolution and learning of language, Functional equations in the theory of dynamic programming. III, Sequentielle Versuchs-Pläne, Gittins' theorem under uncertainty, A Bayesian analysis of human decision-making on bandit problems, Two-armed bandit problem and batch version of the mirror descent algorithm, Randomized Play-the-Leader Rules for Sequential Sampling from Two Populations, An asymptotically optimal strategy for constrained multi-armed bandit problems, Asymptotically optimal algorithms for budgeted multiple play bandits, Small-sample performance of Bernoulli two-armed bandit Bayesian strategies, Stochastic continuum-armed bandits with additive models: minimax regrets and adaptive algorithm, Asymptotically efficient strategies for a stochastic scheduling problem with order constraints., Randomized allocation with nonparametric estimation for a multi-armed bandit problem with covariates, The N-armed bandit with unimodal structure, Asymptotically efficient adaptive allocation rules, Matrices -- compensating the loss of anschauung, Nonparametric bandit methods, Doubly robust policy evaluation and optimization, Multi-armed bandit models for the optimal design of clinical trials: benefits and challenges, RANDOMIZED URN MODELS AND SEQUENTIAL DESIGN, One-armed bandit problem for parallel data processing systems



Cites Work