DocumentCode :
480815
Title :
Learning-Rate Adjusting Q-Learning for Prisoner´s Dilemma Games
Author :
Moriyama, Koichi
Author_Institution :
Inst. of Sci. & Ind. Res., Osaka Univ., Ibaraki
Volume :
2
fYear :
2008
fDate :
9-12 Dec. 2008
Firstpage :
322
Lastpage :
325
Abstract :
Many multiagent Q-learning algorithms have been proposed to date, and most of them aim to converge to a Nash equilibrium, which is not desirable in games like the Prisoner´s Dilemma (PD). In the previous paper, the author proposed the utility-based Q-learning for PD, which used utilities as rewards in order to maintain mutual cooperation once it had occurred. However, since the agent´s action depends on the relation of Q-values the agent has, the mutual cooperation can be maintained by adjusting the learning rate of Q-learning. Thus, in this paper, we deal with the learning rate directly and introduce a new Q-learning method called the learning-rate adjusting Q-learning, or LRA-Q.
Keywords :
game theory; learning (artificial intelligence); multi-agent systems; Nash equilibrium; learning-rate adjusting method; multiagent Q-learning algorithm; mutual cooperation maintenance; prisoner dilemma game; Game theory; Humans; Intelligent agent; Learning; Nash equilibrium; Stochastic processes; Toy industry; Game Theory; Multiagent System; Prisoner´s Dilemma; Reinforcement Learning;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Web Intelligence and Intelligent Agent Technology, 2008. WI-IAT '08. IEEE/WIC/ACM International Conference on
Conference_Location :
Sydney, NSW
Print_ISBN :
978-0-7695-3496-1
Type :
conf
DOI :
10.1109/WIIAT.2008.170
Filename :
4740642
Link To Document :
بازگشت