DocumentCode
463394
Title
Learn to Coordinate with Generic Non-Stationary Opponents
Author
Kaifu, Zhang
Author_Institution
Dept. of Comput. Sci. & Technol., Tsinghua Univ., Beijing
Volume
1
fYear
2006
fDate
17-19 July 2006
Firstpage
558
Lastpage
565
Abstract
Learning to coordinate with non-stationary opponents is a major challenge for adaptive agents. Most previous research investigated only restricted classes of such dynamic opponents. The main contribution of this paper is twofold: (i) A class of generic non-stationary opponents is introduced. The opponents keep mixed strategies which change with less regularity. Its showed that the independent reinforcement learners (ILs), which have neither prior knowledge nor opponent models, cannot coordinate well with this type of opponent, (ii) A new exploration strategy, the DAE (detect and explore) mechanism, is tailored for the ILs in such coordination tasks. This mechanism allows the ILs dynamically detect changes in the opponents behavior and adjust their learning rate and exploration temperature. It´s showed that ILs using this strategy are still able to converge in self-play and are able to coordinate well with the non-stationary opponents
Keywords
learning (artificial intelligence); multi-agent systems; adaptive agents; detect and explore mechanism; dynamic opponents; exploration strategy; generic nonstationary opponents; independent reinforcement learner; Cognitive informatics; Computer science; Decision making; Game theory; Learning; Multiagent systems; Temperature; Testing; Coordination game; Exploration strategy; Non-stationary opponent; Reinforcement learning;
fLanguage
English
Publisher
ieee
Conference_Titel
Cognitive Informatics, 2006. ICCI 2006. 5th IEEE International Conference on
Conference_Location
Beijing
Print_ISBN
1-4244-0475-4
Type
conf
DOI
10.1109/COGINF.2006.365546
Filename
4216463
Link To Document