DocumentCode :
2271683
Title :
Evaluating concurrent reinforcement learners
Author :
Mundhe, Manisha ; Sen, Sandip
Author_Institution :
Dept. of Math. & Comput. Sci., Tulsa Univ., OK, USA
fYear :
2000
fDate :
2000
Firstpage :
421
Lastpage :
422
Abstract :
Assumptions underlying the convergence proofs of reinforcement learning (RL) algorithms like Q-learning are violated when multiple interacting agents adapt their strategies online as a result of learning. Empirical investigations in several domains, however, have produced encouraging results. We evaluate the convergence behavior of concurrent reinforcement learning agents using game matrices as studied by Claus and Boutilier (1998). Variants of simple RL algorithms are evaluated for convergence under increasing number of agents per group, scale up of game matrix size, delayed feedback and game matrix characteristics. Our results show surprising departures from that observed by Claus and Boutilier, particular for larger problem sizes
Keywords :
convergence; learning (artificial intelligence); multi-agent systems; Q-learning; concurrent reinforcement learners; convergence behavior; convergence proofs; delayed feedback; game matrices; multiple interacting agents; Algorithm design and analysis; Convergence; Delay; Feedback; Frequency; Learning; Probability; Utility theory;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
MultiAgent Systems, 2000. Proceedings. Fourth International Conference on
Conference_Location :
Boston, MA
Print_ISBN :
0-7695-0625-9
Type :
conf
DOI :
10.1109/ICMAS.2000.858505
Filename :
858505
Link To Document :
بازگشت