DocumentCode :
2697985
Title :
Distributed neural network-based policy gradient reinforcement learning for multi-robot formations
Author :
Shang, Wen ; Sun, Dong
Author_Institution :
Dept. of Manuf. Eng. & Eng. Manage., City Univ. of Hong Kong, Suzhou
fYear :
2008
fDate :
20-23 June 2008
Firstpage :
113
Lastpage :
118
Abstract :
Multi-robot learning is a challenging task not only because of large and continuous state/action spaces, but also uncertainty and partial observability during learning. This paper presents a distributed policy gradient reinforcement learning (PGRL) methodology of a multi-robot system using neural network as the function approximator. This distributed PGRL algorithm enables each robot to independently decide its policy, which is, however, affected by all the other robots. Neural network is used to generalize over continuous state space as well as discrete/continuous action spaces. A case study on leader-follower formation application is performed to demonstrate the effectiveness of the proposed learning method.
Keywords :
learning (artificial intelligence); multi-robot systems; neural nets; state-space methods; PGRL; continuous state space; distributed neural network; distributed policy gradient reinforcement learning; function approximator; leader-follower formation; multirobot formations; multirobot learning; Function approximation; Learning; Manufacturing automation; Multiagent systems; Multirobot systems; Neural networks; Observability; Orbital robotics; Robots; State-space methods;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Information and Automation, 2008. ICIA 2008. International Conference on
Conference_Location :
Changsha
Print_ISBN :
978-1-4244-2183-1
Electronic_ISBN :
978-1-4244-2184-8
Type :
conf
DOI :
10.1109/ICINFA.2008.4607978
Filename :
4607978
Link To Document :
بازگشت