• DocumentCode
    382902
  • Title

    Approximating the value function for continuous space reinforcement learning in robot control

  • Author

    Buck, Sebastian ; Beetz, Michael ; Schmitt, Thorsten

  • Author_Institution
    Munich Univ. of Technol., Germany
  • Volume
    1
  • fYear
    2002
  • fDate
    2002
  • Firstpage
    1062
  • Abstract
    Many robot learning tasks are very difficult to solve: their state spaces are high dimensional, variables and command parameters are continuously valued, and system states are only partly observable. In this paper, we propose to learn a continuous space value function for reinforcement learning using neural networks trained from data of exploration runs. The learned function is guaranteed to be a lower bound for, and reproduces the characteristic shape of, the accurate value function. We apply our approach to two robot navigation tasks, discuss how to deal with possible problems occurring in practice, and assess its performance.
  • Keywords
    learning (artificial intelligence); mobile robots; neural nets; state-space methods; autonomous robot skills; continuous space value function; neural networks; reinforcement learning; robot learning; robot navigation; state spaces; Acceleration; Learning; Navigation; Neural networks; Orbital robotics; Robot control; Robotics and automation; Shape; State-space methods; Upper bound;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Intelligent Robots and Systems, 2002. IEEE/RSJ International Conference on
  • Print_ISBN
    0-7803-7398-7
  • Type

    conf

  • DOI
    10.1109/IRDS.2002.1041532
  • Filename
    1041532