DocumentCode :
3535212
Title :
Bandits with budgets
Author :
Chong Jiang ; Srikant, R.
Author_Institution :
Dept. of Electr. & Comput. Eng., Univ. of Illinois at Urbana-Champaign, Urbana, IL, USA
fYear :
2013
fDate :
10-13 Dec. 2013
Firstpage :
5345
Lastpage :
5350
Abstract :
Motivated by online advertising applications, we consider a version of the classical multi-armed bandit problem where there is a cost associated with pulling each arm, and a corresponding budget which limits the number of times that an arm can be pulled. We derive regret bounds on the expected reward in such a bandit problem using a modification of the well-known upper confidence bound algorithm UCB1.
Keywords :
advertising; costing; probability; classical multiarmed bandit problem; cost; online advertising applications; upper confidence bound algorithm UCB1; Advertising; Analytical models; Conferences; Google; Random variables; Stochastic processes; Upper bound;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Decision and Control (CDC), 2013 IEEE 52nd Annual Conference on
Conference_Location :
Firenze
ISSN :
0743-1546
Print_ISBN :
978-1-4673-5714-2
Type :
conf
DOI :
10.1109/CDC.2013.6760730
Filename :
6760730
Link To Document :
بازگشت