Title :
Cooperation and coordination between fuzzy reinforcement learning agents in continuous state partially observable Markov decision processes
Author :
Berenji, Hamid R. ; Vengerov, David
Author_Institution :
Div. of Comput. Sci., NASA Ames Res. Center, Moffett Field, CA, USA
Abstract :
We consider a pseudo-realistic world in which one or more opportunities appear and disappear in random locations. Agents use fuzzy reinforcement learning to learn which opportunities are most worthy of pursuing based on their promised rewards, expected lifetimes, path lengths and expected path costs. We show that this world is partially observable because the history of an agent influences the distribution of its future states. We implement a coordination mechanism for allocating opportunities to different agents in the same world. Our results show that optimal team performance results when agents behave in a partially selfish way. We also implement a cooperation mechanism in which agents share experience by using and updating one joint behavior policy. Our results demonstrate that K cooperative agents each learning in a separate world for N time steps outperform K independent agents each learning in a separate world for K*N time steps, with this result becoming more pronounced as the degree of partial observability in the environment increases.
Keywords :
Markov processes; decision theory; fuzzy logic; fuzzy set theory; learning (artificial intelligence); multi-agent systems; observability; Markov decision processes; cooperative systems; coordination mechanism; fuzzy reinforcement learning; multiple agent systems; observability; team performance; Computational intelligence; Costs; Function approximation; Intelligent agent; Intelligent systems; Learning automata; Multiagent systems; NASA; Observability; Tiles;
Conference_Titel :
Fuzzy Systems Conference Proceedings, 1999. FUZZ-IEEE '99. 1999 IEEE International
Conference_Location :
Seoul, South Korea
Print_ISBN :
0-7803-5406-0
DOI :
10.1109/FUZZY.1999.793014