DocumentCode :
2507653
Title :
Learn adaptive dynamic policy under mixed multi-agent environment
Author :
Xiao, Zheng ; Zhang, Shiyong
Author_Institution :
Dept. of Comput.&Inf. Technol., Fudan Univ., Shanghai
fYear :
2008
fDate :
8-11 July 2008
Firstpage :
249
Lastpage :
254
Abstract :
Equilibrium based approach to multi-agent policy learning supposes that all agent uses the same algorithm. Otherwise it is easy for other agents to exploit its policy. Adaptive learning fits well versus fixed or equilibrium policy and in self-play. But it is deficient when against other kinds of adaptive players. On account of this mixed environment, this paper proposes a novel algorithm to learn time dependent policy which can also be capable of adapting to dynamics of policies of others. Policy planning related to time makes a key difference from the learning before which produces stationary policy. Based on two well-known zero-sum games it is demonstrated that agents using this algorithm can get higher utility against some adaptive players, and perform well in self play.
Keywords :
game theory; learning (artificial intelligence); multi-agent systems; planning (artificial intelligence); adaptive learning; equilibrium based approach; multiagent policy learning; policy planning; time dependent policy learning; zero-sum games; Automatic testing; Decision making; Game theory; History; Information technology; Learning; Nash equilibrium; Probability distribution; Stochastic processes; Transfer functions;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Computer and Information Technology, 2008. CIT 2008. 8th IEEE International Conference on
Conference_Location :
Sydney, NSW
Print_ISBN :
978-1-4244-2357-6
Electronic_ISBN :
978-1-4244-2358-3
Type :
conf
DOI :
10.1109/CIT.2008.4594682
Filename :
4594682
Link To Document :
بازگشت