Title :
Optimal strategies for families of alternative bandit processes
Author :
Glazebrook, K.D.
Author_Institution :
University of Newcastle upon Tyne, Newcastle upon Tyne, UK
fDate :
8/1/1983 12:00:00 AM
Abstract :
Many stochastic resource allocation problems may be formulated as families of alternative bandit processes. One example is the classical one-armed bandit problem recently studied by Kumar and Seidman. Optimal strategies for such problems are known to be determined by a collection of dynamic allocation indexes (DAI´s). The aim of this note is to bring this important result to the attention of control theorists and to give a new proof of it. Applications and some related work are also discussed.
Keywords :
Adaptive control; Markov processes; Optimal stochastic control; Resource management; Sequential decision procedures; Stochastic optimal control; Adaptive control; Costs; Dynamic programming; Equations; Resource management; Silicon compounds; State-space methods; Stochastic processes; Writing;
Journal_Title :
Automatic Control, IEEE Transactions on
DOI :
10.1109/TAC.1983.1103327