Title :
A collaborative distributed multi-agent reinforcement learning technique for dynamic agent shortest path planning via selected sub-goals in complex cluttered environments
Author :
Megherbi, D.B. ; Kim, M.
Author_Institution :
Dept. of Electr. & Comput. Eng., Univ. of Massachusetts, Lowell, MA, USA
Abstract :
Collaborative monitoring of large infrastructures, such as military, transportation and maritime systems are decisive issues in many surveillance, protection, and security applications. In many of these applications, dynamic multi-agent systems using reinforcement learning for agents´ autonomous path planning, where agents could be moving randomly to reach their respective goals and avoiding topographical obstacles intelligently, becomes a challenging problem. This is specially so in a dynamic agent environment. In our prior work we presented an intelligent multi-agent hybrid reactive and reinforcement learning technique for collaborative autonomous agent path planning for monitoring Critical Key Infrastructures and Resources (CKIR) in a geographically and a computationally distributed systems. Here agent monitoring of large environments is reduced to monitoring of relatively smaller track-able geographically distributed agent environment regions. In this paper we tackle this problem in the challenging case of complex and cluttered environments, where agents´ initial random-walk paths become challenging and relatively nonconverging. Here we propose a multi-agent distributed hybrid reactive re-enforcement learning technique based on selected agent intermediary sub-goals using a learning reward scheme in a distributed-computing memory setting. Various case study scenarios are presented for convergence study to the shortest minimum-amount-of-time exploratory steps for faster and efficient agent learning. In this work the distributed dynamic agent communication is done via a Message Passing Interface (MPI).
Keywords :
application program interfaces; computerised monitoring; critical infrastructures; groupware; learning (artificial intelligence); message passing; multi-agent systems; path planning; CKIR; MPI; agent initial random-walk path; collaborative autonomous agent path planning; collaborative distributed multiagent reinforcement learning technique; complex cluttered environments; critical key infrastructures and resources monitoring; distributed dynamic agent communication; distributed-computing memory setting; dynamic agent shortest path planning; dynamic multiagent systems; geographically distributed agent environment regions; intelligent multiagent hybrid reactive reinforcement learning technique; large infrastructure collaborative monitoring; learning reward scheme; message passing interface; shortest minimum-amount-of-time exploratory steps; subgoal selecion; topographical obstacle avoidance; Computer architecture; Conferences; Learning (artificial intelligence); Monitoring; Multi-agent systems; Path planning; Peer-to-peer computing; Intelligent multi-agent systems; Key infrastructures and resources; Machine learning; distributed systems and networks; sub-goals; tracking of friendly and enemy targets;
Conference_Titel :
Cognitive Methods in Situation Awareness and Decision Support (CogSIMA), 2015 IEEE International Inter-Disciplinary Conference on
Conference_Location :
Orlando, FL
DOI :
10.1109/COGSIMA.2015.7108185