DocumentCode
3383072
Title
Greedy exploration policy of Q-learning based on state balance
Author
Zheng, Yu ; Luo, Siwei ; Zhang, Jing
Author_Institution
Sch. of Comput. & Inf. Technol., Beijing Jiaotong Univ., Beijing
fYear
2005
fDate
21-24 Nov. 2005
Firstpage
1
Lastpage
4
Abstract
Q-learning is one of the successfully established algorithms for the reinforcement learning, which has been widely used to the intelligent control system, such as the control of robot pose. However, curse of dimensionality and difficulty in convergence exist in Q-learning arising from random exploration policy. In this paper, we propose a greedy exploration policy of Q-learning with rule guidance. This exploration policy can reduce the non-optimal action exploration as more as possible, and speed up the convergence of Q-learning. Simulation results indicate the effectiveness of the proposed method.
Keywords
learning (artificial intelligence); Q-learning; greedy exploration policy; intelligent control system; nonoptimal action exploration; reinforcement learning; rule guidance; state balance; Acceleration; Computational modeling; Control systems; Electronic mail; Information technology; Learning; Optimal control; Robot control; State estimation; State-space methods;
fLanguage
English
Publisher
ieee
Conference_Titel
TENCON 2005 2005 IEEE Region 10
Conference_Location
Melbourne, Qld.
Print_ISBN
0-7803-9311-2
Electronic_ISBN
0-7803-9312-0
Type
conf
DOI
10.1109/TENCON.2005.300987
Filename
4085232
Link To Document