DocumentCode :
50456
Title :
Accelerating Multiagent Reinforcement Learning by Equilibrium Transfer
Author :
Yujing Hu ; Yang Gao ; Bo An
Author_Institution :
Dept. of Comput. Sci., Nanjing Univ., Nanjing, China
Volume :
45
Issue :
7
fYear :
2015
fDate :
Jul-15
Firstpage :
1289
Lastpage :
1302
Abstract :
An important approach in multiagent reinforcement learning (MARL) is equilibrium-based MARL, which adopts equilibrium solution concepts in game theory and requires agents to play equilibrium strategies at each state. However, most existing equilibrium-based MARL algorithms cannot scale due to a large number of computationally expensive equilibrium computations (e.g., computing Nash equilibria is PPAD-hard) during learning. For the first time, this paper finds that during the learning process of equilibrium-based MARL, the one-shot games corresponding to each state´s successive visits often have the same or similar equilibria (for some states more than 90% of games corresponding to successive visits have similar equilibria). Inspired by this observation, this paper proposes to use equilibrium transfer to accelerate equilibrium-based MARL. The key idea of equilibrium transfer is to reuse previously computed equilibria when each agent has a small incentive to deviate. By introducing transfer loss and transfer condition, a novel framework called equilibrium transfer-based MARL is proposed. We prove that although equilibrium transfer brings transfer loss, equilibrium-based MARL algorithms can still converge to an equilibrium policy under certain assumptions. Experimental results in widely used benchmarks (e.g., grid world game, soccer game, and wall game) show that the proposed framework: 1) not only significantly accelerates equilibrium-based MARL (up to 96.7% reduction in learning time), but also achieves higher average rewards than algorithms without equilibrium transfer and 2) scales significantly better than algorithms without equilibrium transfer when the state/action space grows and the number of agents increases.
Keywords :
game theory; learning (artificial intelligence); multi-agent systems; equilibrium solution concept; equilibrium transfer; equilibrium-based MARL; game theory; multiagent reinforcement learning acceleration; one-shot games; state successive visit; state/action space; transfer condition; transfer loss; Algorithm design and analysis; Game theory; Games; Joints; Learning (artificial intelligence); Loss measurement; Markov processes; Equilibrium; equilibrium transfer; game theory; multiagent reinforcement learning (MARL); multiagent reinforcement learning (MARL).;
fLanguage :
English
Journal_Title :
Cybernetics, IEEE Transactions on
Publisher :
ieee
ISSN :
2168-2267
Type :
jour
DOI :
10.1109/TCYB.2014.2349152
Filename :
6888505
Link To Document :
بازگشت