• DocumentCode
    2697985
  • Title

    Distributed neural network-based policy gradient reinforcement learning for multi-robot formations

  • Author

    Shang, Wen ; Sun, Dong

  • Author_Institution
    Dept. of Manuf. Eng. & Eng. Manage., City Univ. of Hong Kong, Suzhou
  • fYear
    2008
  • fDate
    20-23 June 2008
  • Firstpage
    113
  • Lastpage
    118
  • Abstract
    Multi-robot learning is a challenging task not only because of large and continuous state/action spaces, but also uncertainty and partial observability during learning. This paper presents a distributed policy gradient reinforcement learning (PGRL) methodology of a multi-robot system using neural network as the function approximator. This distributed PGRL algorithm enables each robot to independently decide its policy, which is, however, affected by all the other robots. Neural network is used to generalize over continuous state space as well as discrete/continuous action spaces. A case study on leader-follower formation application is performed to demonstrate the effectiveness of the proposed learning method.
  • Keywords
    learning (artificial intelligence); multi-robot systems; neural nets; state-space methods; PGRL; continuous state space; distributed neural network; distributed policy gradient reinforcement learning; function approximator; leader-follower formation; multirobot formations; multirobot learning; Function approximation; Learning; Manufacturing automation; Multiagent systems; Multirobot systems; Neural networks; Observability; Orbital robotics; Robots; State-space methods;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Information and Automation, 2008. ICIA 2008. International Conference on
  • Conference_Location
    Changsha
  • Print_ISBN
    978-1-4244-2183-1
  • Electronic_ISBN
    978-1-4244-2184-8
  • Type

    conf

  • DOI
    10.1109/ICINFA.2008.4607978
  • Filename
    4607978