Title :
Distributed dynamic reinforcement of efficient outcomes in multiagent coordination
Author :
Chasparis, Georgios C. ; Shamma, Jeff S.
Author_Institution :
Dept. of Mech. & Aerosp. Eng., Univ. of California Los Angeles, Los Angeles, CA, USA
Abstract :
We consider the problem of achieving distributed convergence to coordination in a multiagent environment. Each agent is modeled as a learning automaton which repeatedly interacts with an unknown environment, receives a reward, and updates the probabilities of its next action based on its own previous actions and received rewards. In this class of problems, more than one stable equilibrium (i.e., coordination structure) exists. We analyze the dynamic behavior of the distributed system in terms of convergence to an efficient equilibrium, suitably defined. In particular, we analyze the effect of dynamic processing on convergence properties, where agents include the derivative of their own reward into the decision process (i.e., derivative action). We show that derivative action can be used as an equilibrium selection scheme by appropriately adjusting derivative feedback gains.
Keywords :
convergence; distributed processing; learning automata; multi-agent systems; convergence properties; decision process; derivative action; derivative feedback gains; distributed convergence; distributed dynamic reinforcement; distributed system; efficient equilibrium; equilibrium selection scheme; learning automaton; multiagent coordination; Asymptotic stability; Convergence; Games; Heuristic algorithms; Learning (artificial intelligence); Learning automata; Vectors;
Conference_Titel :
Control Conference (ECC), 2007 European
Conference_Location :
Kos
Print_ISBN :
978-3-9524173-8-6