• DocumentCode
    138707
  • Title

    Prior-free and prior-dependent regret bounds for Thompson Sampling

  • Author

    Bubeck, Sebastian ; Che-Yu Liu

  • Author_Institution
    Dept. of Oper. Res. & Financial Eng., Princeton Univ., Princeton, NJ, USA
  • fYear
    2014
  • fDate
    19-21 March 2014
  • Firstpage
    1
  • Lastpage
    9
  • Abstract
    We consider the stochastic multi-armed bandit problem with a prior distribution on the reward distributions. We are interested in studying prior-free and prior-dependent regret bounds, very much in the same spirit than the usual distribution-free and distribution-dependent bounds for the non-Bayesian stochastic bandit. We first show that Thompson Sampling attains an optimal prior-free bound in the sense that for any prior distribution its Bayesian regret is bounded from above by 14√nK. This result is unimprovable in the sense that there exists a prior distribution such that any algorithm has a Bayesian regret bounded from below by 1 over 20√nK. We also study the case of priors for the setting of Bubeck et al. [2013] (where the optimal mean is known as well as a lower bound on the smallest gap) and we show that in this case the regret of Thompson Sampling is in fact uniformly bounded over time, thus showing that Thompson Sampling can greatly take advantage of the nice properties of these priors.
  • Keywords
    Bayes methods; integral equations; stochastic processes; Bayesian regret; Thompson sampling; nonBayesian stochastic bandit; optimal prior-free bound; prior distribution; prior-dependent regret bounds; prior-free regret bounds; reward distributions; stochastic multiarmed bandit problem; Abstracts; Business process re-engineering;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Information Sciences and Systems (CISS), 2014 48th Annual Conference on
  • Conference_Location
    Princeton, NJ
  • Type

    conf

  • DOI
    10.1109/CISS.2014.6814158
  • Filename
    6814158