Title :
Extend Single-agent Reinforcement Learning Approach to a Multi-robot Cooperative Task in an Unknown Dynamic Environment
Author :
Wang, Ying ; De Silva, Clarence W.
Author_Institution :
British Columbia Univ., Vancouver
Abstract :
Machine learning technology helps multi-robot systems to carry out desired tasks in an unknown dynamic environment. In this paper, we extend the single-agent Q-learning algorithm to a multi-robot box-pushing system in an unknown dynamic environment with random obstacle distribution. There are two kinds of extensions available: directly extending MDP (Markov decision process) based Q-learning to the multi-robot domain, and SG-based (stochastic game based) Q-learning. Here, we select the first kind of extension because of its simplicity. The learning space, the box dynamics, and the reward function etc. are presented in this paper. Furthermore, a simulation system is developed and its results show effectiveness, robustness and adaptivity of this learning-based multi-robot system. Our statistical analysis of the results also shows that the robots learned correct cooperative strategy even in a dynamic environment.
Keywords :
Markov processes; collision avoidance; cooperative systems; decision theory; game theory; intelligent robots; learning (artificial intelligence); multi-robot systems; Markov decision process; box-pushing system; multirobot cooperative task; random obstacle distribution; single-agent reinforcement learning approach; statistical analysis; stochastic game based Q-learning; unknown dynamic environment; Algorithm design and analysis; Game theory; Machine learning; Orbital robotics; Paper technology; Predictive models; Robots; Robustness; Solid modeling; Working environment noise;
Conference_Titel :
Neural Networks, 2006. IJCNN '06. International Joint Conference on
Conference_Location :
Vancouver, BC
Print_ISBN :
0-7803-9490-9
DOI :
10.1109/IJCNN.2006.247204