DocumentCode :
2473071
Title :
Adaptive state aggregation for reinforcement learning
Author :
Hwang, Kao-Shing ; Chen, Yu-Jen ; Jiang, Wei-Cheng
Author_Institution :
Dept. of Electr. Eng., Nat. Sun Yat-Sen Univ., Kaohsiung, Taiwan
fYear :
2012
fDate :
14-17 Oct. 2012
Firstpage :
2452
Lastpage :
2456
Abstract :
State partition is an important issue in reinforcement learning, because it has a significant effect on the performance. In this paper, an adaptive state partition method is presented for discretizing the state space adaptively and makes use of decision trees effectively. The proposed method splits the state space according to the temporal difference generated by the reinforcement learning. Consequently, the reinforcement learning uses the state space partitioned by the decision tree to learn the policy simultaneously. For avoiding a trivial partition, sibling nodes are pruned according to the Activity and the Reliability. A Monte-Carlo Tree Search (MCTS) is also proposed to explore the policy. A simulation for approaching goal has been conducted to demonstrate that the proposed method can achieve the design goal.
Keywords :
Monte Carlo methods; decision trees; learning (artificial intelligence); state-space methods; tree searching; MCTS; Monte-Carlo tree search; adaptive state aggregation; adaptive state partition method; decision trees; reinforcement learning; reliability; sibling nodes; state space; temporal difference; trivial partition; Decision trees; Estimation error; Learning; Markov processes; Monte Carlo methods; Reliability; Vectors; MCTS; decision tree; reinforcement learning;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Systems, Man, and Cybernetics (SMC), 2012 IEEE International Conference on
Conference_Location :
Seoul
Print_ISBN :
978-1-4673-1713-9
Electronic_ISBN :
978-1-4673-1712-2
Type :
conf
DOI :
10.1109/ICSMC.2012.6378111
Filename :
6378111
Link To Document :
بازگشت