DocumentCode
709154
Title
A collaborative distributed multi-agent reinforcement learning technique for dynamic agent shortest path planning via selected sub-goals in complex cluttered environments
Author
Megherbi, D.B. ; Kim, M.
Author_Institution
Dept. of Electr. & Comput. Eng., Univ. of Massachusetts, Lowell, MA, USA
fYear
2015
fDate
9-12 March 2015
Firstpage
118
Lastpage
124
Abstract
Collaborative monitoring of large infrastructures, such as military, transportation and maritime systems are decisive issues in many surveillance, protection, and security applications. In many of these applications, dynamic multi-agent systems using reinforcement learning for agents´ autonomous path planning, where agents could be moving randomly to reach their respective goals and avoiding topographical obstacles intelligently, becomes a challenging problem. This is specially so in a dynamic agent environment. In our prior work we presented an intelligent multi-agent hybrid reactive and reinforcement learning technique for collaborative autonomous agent path planning for monitoring Critical Key Infrastructures and Resources (CKIR) in a geographically and a computationally distributed systems. Here agent monitoring of large environments is reduced to monitoring of relatively smaller track-able geographically distributed agent environment regions. In this paper we tackle this problem in the challenging case of complex and cluttered environments, where agents´ initial random-walk paths become challenging and relatively nonconverging. Here we propose a multi-agent distributed hybrid reactive re-enforcement learning technique based on selected agent intermediary sub-goals using a learning reward scheme in a distributed-computing memory setting. Various case study scenarios are presented for convergence study to the shortest minimum-amount-of-time exploratory steps for faster and efficient agent learning. In this work the distributed dynamic agent communication is done via a Message Passing Interface (MPI).
Keywords
application program interfaces; computerised monitoring; critical infrastructures; groupware; learning (artificial intelligence); message passing; multi-agent systems; path planning; CKIR; MPI; agent initial random-walk path; collaborative autonomous agent path planning; collaborative distributed multiagent reinforcement learning technique; complex cluttered environments; critical key infrastructures and resources monitoring; distributed dynamic agent communication; distributed-computing memory setting; dynamic agent shortest path planning; dynamic multiagent systems; geographically distributed agent environment regions; intelligent multiagent hybrid reactive reinforcement learning technique; large infrastructure collaborative monitoring; learning reward scheme; message passing interface; shortest minimum-amount-of-time exploratory steps; subgoal selecion; topographical obstacle avoidance; Computer architecture; Conferences; Learning (artificial intelligence); Monitoring; Multi-agent systems; Path planning; Peer-to-peer computing; Intelligent multi-agent systems; Key infrastructures and resources; Machine learning; distributed systems and networks; sub-goals; tracking of friendly and enemy targets;
fLanguage
English
Publisher
ieee
Conference_Titel
Cognitive Methods in Situation Awareness and Decision Support (CogSIMA), 2015 IEEE International Inter-Disciplinary Conference on
Conference_Location
Orlando, FL
Type
conf
DOI
10.1109/COGSIMA.2015.7108185
Filename
7108185
Link To Document