scientific article; zbMATH DE number 3638998

From MaRDI portal

zbMath0411.62055MaRDI QIDQ4197923

J. C. Gittins

Publication date: 1979


Title: zbMATH Open Web Interface contents unavailable due to conflicting licenses.



Related Items

Optimal learning and experimentation in bandit problems., An optimal stopping time problem with time average cost in a bounded interval, Fast and slow enigmas and parental guidance, SEH: size estimate hedging for single-server queues, The performance of forwards induction policies, Parallel search for the best alternative, Approximation algorithms for stochastic combinatorial optimization problems, Sequencing unreliable jobs on parallel machines, Selecting among scheduled projects, Stochastic scheduling and forwards induction, Learning theorem proving components, Infomax strategies for an optimal balance between exploration and exploitation, Multi-armed bandits with simple arms, Search and active learning with correlated information: empirical evidence from mid-Atlantic clam fishermen, Bandit and covariate processes, with finite or non-denumerable set of arms, Response-adaptive designs for clinical trials: simultaneous learning from multiple patients, Competing Markov decision processes, Robust experimentation in the continuous time bandit problem, Multi-armed bandit processes with optimal selection of the operating times, On Gittins' index theorem in continuous time, Marginal productivity index policies for scheduling a multiclass delay-/loss-sensitive queue, Resource capacity allocation to stochastic dynamic competitors: knapsack problem for perishable items and index-knapsack heuristic, Four proofs of Gittins' multiarmed bandit theorem, Optimal decision indices for R\&D project evaluation in the pharmaceutical industry: Pearson index versus Gittins index, Control, cost, and confidence: perseverance and procrastination in the face of failure, A Bayesian adaptive design for clinical trials in rare diseases, Assessing the effects of machine breakdowns in stochastic scheduling, Bounds on optimal values in stochastic scheduling, Selecting jobs for scheduling on a machine subject to failure, Kullback-Leibler upper confidence bounds for optimal sequential allocation, A fluid approach to large volume job shop scheduling, Stochastic scheduling on a single machine subject to multiple breakdowns according to different probabilities, The multi-armed bandit problem: an efficient nonparametric solution, The multi-armed bandit, with constraints, Derman's book as inspiration: some results on LP for MDPs, Multi-machine preventive maintenance scheduling with imperfect interventions: a restless bandit approach, The archievable region method in the optimal control of queueing systems; formulations, bounds and policies, The expected asymptotical ratio for preemptive stochastic online problem, General time consistent discounting, Rational status quo, Optimal experimental design for a class of bandit problems, An application of Edgeworth expansion in Bayesian inferences: Optimal sample sizes in clinical trials, Truthful learning mechanisms for multi-slot sponsored search auctions with externalities, Bernoulli two-armed bandits with geometric termination, On the smoothness of value functions and the existence of optimal strategies in diffusion models, Algorithms and mechanisms for procuring services with uncertain durations using redundancy, A Bayesian approach to the triage problem with imperfect classification, Decomposing risk in an exploitation-exploration problem with endogenous termination time, Optimal Bayesian strategies for the infinite-armed Bernoulli bandit, On the evaluation of fixed permutations as strategies in stochastic scheduling, Branching bandits: A sequential search process with correlated pay-offs., Regret bounds for sleeping experts and bandits, Algorithms for evaluating the dynamic allocation index, On optimal search with unknown detection probabilities, A conservative index heuristic for routing problems with multiple heterogeneous service facilities, Approximation results in parallel machines stochastic scheduling, A unified framework for stochastic optimization, Sequential process control under capacity constraints., Regret bounds for restless Markov bandits, Dynamic priority allocation via restless bandit marginal productivity indices, A novel scheduling index rule proposal for QoE maximization in wireless networks, Optimal learning before choice, On the evaluation of strategies for branching bandit processes, Dynamic survival bias in optimal stopping problems, Stochastic scheduling: a short history of index policies and new approaches to index generation for dynamic resource allocation, A generalized Gittins index for a Markov chain and its recursive calculation, On Bayesian index policies for sequential resource allocation, Arbitrary side observations in bandit problems, Optimal stopping for Brownian motion with applications to sequential analysis and option pricing, Randomized allocation with nonparametric estimation for contextual multi-armed bandits with delayed rewards, General notions of indexability for queueing control and asset management, On the optimal amount of experimentation in sequential decision problems, Using adaptive learning in credit scoring to estimate take-up probability distribution, Index policies for the maintenance of a collection of machines by a set of repairmen, Player-compatible learning and player-compatible equilibrium, Some best possible results for a discounted one armed bandit, Boundary crossing probabilities for general exponential families, A survey of Markov decision models for control of networks of queues, On index policies for stochastic minsum scheduling, Optimal learning for sequential sampling with non-parametric beliefs, Optimal learning with a local parametric belief model, On the dynamic allocation of assets subject to failure, A model of experimentation with information externalities, Risk aversion in expected intertemporal discounted utilities bandit problems, Gittins' theorem under uncertainty, Learning, risk attitude and hot stoves in restless bandit problems, A Bayesian analysis of human decision-making on bandit problems, Dismemberment and design for controlling the replication variance of regret for the multi-armed bandit, Asymptotically optimal algorithms for budgeted multiple play bandits, On scheduling influential stochastic tasks on a single machine, Matrices -- compensating the loss of anschauung, On the almost sure convergence of adaptive allocation procedures, Discrete multiarmed bandits and multiparameter processes, Multi-armed bandit models for the optimal design of clinical trials: benefits and challenges, Optimal discrete search with technological choice, A job shop scheduling heuristic for varying reward structures, Dynamic decision making for graphical models applied to oil exploration, Algorithmic aspects of mean-variance optimization in Markov decision processes, Optimal stopping problems for multiarmed bandit processes with arms' independence, Experimentation and project selection: screening and learning, Functional Sequential Treatment Allocation, A forwards induction approach to candidate drug selection, Conditions for indexability of restless bandits and an algorithm to compute Whittle index, Optimistic Gittins Indices, On the optimal allocation of service to impatient tasks, Multi-Actor Markov Decision Processes, MULTI-ARMED BANDITS WITH COVARIATES:THEORY AND APPLICATIONS, Bayesian Exploration: Incentivizing Exploration in Bayesian Games, Whittle’s Index Policy for Multi-Target Tracking with Jamming and Nondetections, Generalized Restless Bandits and the Knapsack Problem for Perishable Inventories, Dynamic Learning and Decision Making via Basis Weight Vectors, Optimal activation of halting multi‐armed bandit models, A novel statistical test for treatment differences in clinical trials using a response‐adaptive forward‐looking Gittins Index Rule, Encounters with Martingales in Statistics and Stochastic Optimization, Multi-armed bandit problem with online clustering as side information, Testing indexability and computing Whittle and Gittins index in subcubic time, MULTI-ARMED BANDITS UNDER GENERAL DEPRECIATION AND COMMITMENT, INDEXABILITY AND OPTIMAL INDEX POLICIES FOR A CLASS OF REINITIALISING RESTLESS BANDITS, Treatment recommendation with distributional targets, Exponential asymptotic optimality of Whittle index policy, Empirical Gittins index strategies with \(\varepsilon\)-explorations for multi-armed bandit problems, A general approximation method for optimal stopping and random delay, Stochastic Probing with Increasing Precision, Optimal Exploration–Exploitation in a Multi-armed Bandit Problem with Non-stationary Rewards, Index policies for discounted bandit problems with availability constraints, Learning the distribution with largest mean: two bandit frameworks, A Verification Theorem for Threshold-Indexability of Real-State Discounted Restless Bandits, Finite-Time Analysis for the Knowledge-Gradient Policy, BANDIT STRATEGIES EVALUATED IN THE CONTEXT OF CLINICAL TRIALS IN RARE LIFE-THREATENING DISEASES, Coping with Incomplete Information in Scheduling — Stochastic and Online Models, Nonstationary Bandits with Habituation and Recovery Dynamics, Learning in Combinatorial Optimization: What and How to Explore, Simple Bayesian Algorithms for Best-Arm Identification, Some indexable families of restless bandit problems, An Approximation Approach for Response-Adaptive Clinical Trial Design, Approximate Dynamic Programming based on High Dimensional Model Representation, Monotone Policies and Indexability for Bidirectional Restless Bandits, Explicit Gittins Indices for a Class of Superdiffusive Processes, Tax problems in the undiscounted case, Survey of linear programming for standard and nonstandard Markovian control problems. Part II: Applications, Two-Armed Restless Bandits with Imperfect Information: Stochastic Control and Indexability, When to Abandon a Research Project and Search for a New One, Bayesian Incentive-Compatible Bandit Exploration, Spinning plates and squad systems: policies for bi-directional restless bandits, Un ordonnancement dynamique de tâches stochastiques sur un seul processeur, Stochastic graph exploration, Unnamed Item, Online Collaborative Filtering on Graphs, A Continuous-Time Markov Decision Process for Infrastructure Surveillance, Generalized Bandit Problems, Learning Unknown Service Rates in Queues: A Multiarmed Bandit Approach, A Tight 2-Approximation for Preemptive Stochastic Scheduling, A simulation-based approach to stochastic dynamic programming, Time-Constrained Restless Bandits and the Knapsack Problem for Perishable Items (Extended Abstract), A General Theory of MultiArmed Bandit Processes with Constrained Arm Switches, A Bandit-Learning Approach to Multifidelity Approximation, A Restless Bandit Model for Resource Allocation, Competition, and Reservation