Title : 
Learning-Rate Adjusting Q-Learning for Prisoner´s Dilemma Games
         
        
            Author : 
Moriyama, Koichi
         
        
            Author_Institution : 
Inst. of Sci. & Ind. Res., Osaka Univ., Ibaraki
         
        
        
        
        
        
        
            Abstract : 
Many multiagent Q-learning algorithms have been proposed to date, and most of them aim to converge to a Nash equilibrium, which is not desirable in games like the Prisoner´s Dilemma (PD). In the previous paper, the author proposed the utility-based Q-learning for PD, which used utilities as rewards in order to maintain mutual cooperation once it had occurred. However, since the agent´s action depends on the relation of Q-values the agent has, the mutual cooperation can be maintained by adjusting the learning rate of Q-learning. Thus, in this paper, we deal with the learning rate directly and introduce a new Q-learning method called the learning-rate adjusting Q-learning, or LRA-Q.
         
        
            Keywords : 
game theory; learning (artificial intelligence); multi-agent systems; Nash equilibrium; learning-rate adjusting method; multiagent Q-learning algorithm; mutual cooperation maintenance; prisoner dilemma game; Game theory; Humans; Intelligent agent; Learning; Nash equilibrium; Stochastic processes; Toy industry; Game Theory; Multiagent System; Prisoner´s Dilemma; Reinforcement Learning;
         
        
        
        
            Conference_Titel : 
Web Intelligence and Intelligent Agent Technology, 2008. WI-IAT '08. IEEE/WIC/ACM International Conference on
         
        
            Conference_Location : 
Sydney, NSW
         
        
            Print_ISBN : 
978-0-7695-3496-1
         
        
        
            DOI : 
10.1109/WIIAT.2008.170