• DocumentCode
    3267971
  • Title

    Caterpillar robot locomotion based on Q-Learning using objective/subjective reward

  • Author

    Yamashina, Ryota ; Kuroda, Masafumi ; Yabuta, Tetsuro

  • Author_Institution
    Dept. of Mech. Eng. & Mater. Sci., Yokohama Nat. Univ., Yokohama, Japan
  • fYear
    2011
  • fDate
    20-22 Dec. 2011
  • Firstpage
    1311
  • Lastpage
    1316
  • Abstract
    This paper presents an application of reinforcement learning, an unsupervised learning method, to a biological robot. This study focused on the primitive forward motion of a caterpillar robot to reveal how the robot obtains an optimal motion form in the Q-Learning process. First, this paper verifies that Q-Learning allows the caterpillar robot to move in the forward direction and examines the evolutionary process. Next, this paper discusses the emergence of the motion form using objective rewards. Finally, it examines the emergence of the motion using Q-Learning with subjective rewards in order to clarify the difference between the learning results. This examination provides a novel perspective on human robot interaction (HRI) via reinforcement learning.
  • Keywords
    control engineering computing; human-robot interaction; learning (artificial intelligence); mobile robots; path planning; Q-learning; biological robot; caterpillar robot locomotion; evolutionary process; human robot interaction; objective reward; optimal motion; primitive forward motion; reinforcement learning; subjective reward; unsupervised learning method; Convergence; Databases; Humans; Learning; Robot sensing systems; Silicon;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    System Integration (SII), 2011 IEEE/SICE International Symposium on
  • Conference_Location
    Kyoto
  • Print_ISBN
    978-1-4577-1523-5
  • Type

    conf

  • DOI
    10.1109/SII.2011.6147638
  • Filename
    6147638