• DocumentCode
    607751
  • Title

    Sequential Monte Carlo samplers for model-based reinforcement learning

  • Author

    Sonmez, O. ; Cemgil, A.T.

  • Author_Institution
    Bilgisayar Muhendisligi Bolumu, Bogazici Univ., Istanbul, Turkey
  • fYear
    2013
  • fDate
    24-26 April 2013
  • Firstpage
    1
  • Lastpage
    4
  • Abstract
    Reinforcement learning problems are generally solved by using fixed-point iterations that converge to the suboptimal solutions of Bellman equations. However, it is also possible to formalize this problem as an equivalent likelihood maximization problem and employ probabilistic inference methods. We proposed an expectation-maximization algorithm that utilizes sequential Monte Carlo samplers with Metropolis-Hastings kernels in its expectation step to solve the model-based version. Then, we evaluate our algorithm on mountain-car problem which is a benchmark reinforcement learning problem.
  • Keywords
    Monte Carlo methods; expectation-maximisation algorithm; learning (artificial intelligence); Bellman equations; equivalent likelihood maximization problem; expectation-maximization algorithm; fixed-point iterations; model-based reinforcement learning; probabilistic inference methods; sequential Monte Carlo samplers; Electronic mail; Learning (artificial intelligence); Markov processes; Mathematical model; Monte Carlo methods; Presses; Probabilistic logic; Expectation-Maximization; Markov Decision Processes; Metropolis-Hastings; Reinforcement Learning; Sequential Monte Carlo Samplers;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Signal Processing and Communications Applications Conference (SIU), 2013 21st
  • Conference_Location
    Haspolat
  • Print_ISBN
    978-1-4673-5562-9
  • Electronic_ISBN
    978-1-4673-5561-2
  • Type

    conf

  • DOI
    10.1109/SIU.2013.6531412
  • Filename
    6531412