DocumentCode :
453874
Title :
Option Discovery in Reinforcement Learning using Frequent Common Subsequences of Actions
Author :
Girgin, Sertan ; Polat, Faruk
Author_Institution :
Dept. of Comput. Eng., Middle East Tech. Univ., Ankara
Volume :
1
fYear :
2005
fDate :
28-30 Nov. 2005
Firstpage :
371
Lastpage :
376
Abstract :
Temporally abstract actions, or options, facilitate learning in large and complex domains by exploiting sub-tasks and hierarchical structure of the problem formed by these sub-tasks. In this paper, we study automatic generation of options using common sub-sequences derived from the state transition histories collected as learning progresses. The standard Q-learning algorithm is extended to use generated options transparently, and effectiveness of the method is demonstrated in Dietterich´s Taxi domain
Keywords :
Markov processes; learning (artificial intelligence); set theory; Dietterich Taxi domain; frequent common subsequence; option discovery; reinforcement learning; standard Q-learning algorithm; temporally abstract action; Acceleration; Clustering algorithms; Computational intelligence; Computational modeling; History; Joining processes; Learning; Partitioning algorithms; State-space methods; Topology;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Computational Intelligence for Modelling, Control and Automation, 2005 and International Conference on Intelligent Agents, Web Technologies and Internet Commerce, International Conference on
Conference_Location :
Vienna
Print_ISBN :
0-7695-2504-0
Type :
conf
DOI :
10.1109/CIMCA.2005.1631294
Filename :
1631294
Link To Document :
بازگشت