DocumentCode :
842715
Title :
Optimal strategies for families of alternative bandit processes
Author :
Glazebrook, K.D.
Author_Institution :
University of Newcastle upon Tyne, Newcastle upon Tyne, UK
Volume :
28
Issue :
8
fYear :
1983
fDate :
8/1/1983 12:00:00 AM
Firstpage :
858
Lastpage :
861
Abstract :
Many stochastic resource allocation problems may be formulated as families of alternative bandit processes. One example is the classical one-armed bandit problem recently studied by Kumar and Seidman. Optimal strategies for such problems are known to be determined by a collection of dynamic allocation indexes (DAI´s). The aim of this note is to bring this important result to the attention of control theorists and to give a new proof of it. Applications and some related work are also discussed.
Keywords :
Adaptive control; Markov processes; Optimal stochastic control; Resource management; Sequential decision procedures; Stochastic optimal control; Adaptive control; Costs; Dynamic programming; Equations; Resource management; Silicon compounds; State-space methods; Stochastic processes; Writing;
fLanguage :
English
Journal_Title :
Automatic Control, IEEE Transactions on
Publisher :
ieee
ISSN :
0018-9286
Type :
jour
DOI :
10.1109/TAC.1983.1103327
Filename :
1103327
Link To Document :
بازگشت