DocumentCode
1752965
Title
An Improved Q-learning Algorithm Based on Exploration Region Expansion Strategy
Author
Gao, Qingji ; Hong, Bingrong ; He, Zhendong ; Liu, Jie ; Niu, Guochen
Author_Institution
Dept. of Comput. Sci. & Technol., Harbin Inst. of Technol.
Volume
1
fYear
0
fDate
0-0 0
Firstpage
4167
Lastpage
4170
Abstract
In order to find a good solution to one of the key problems in Q-learning algorithm - keeping the balance between exploration and exploitation, an improved Q-learning algorithm based on exploration region expansion strategy is proposed on the base of Metropolis criterion-based Q-learning. With this strategy, the exploration blindness in the entire environment is eliminated, and the learning efficiency is increased. Meanwhile, other feasible path is sought where agent encounters obstacles, which makes the implementation of the algorithm on real robot easy. An automatic termination condition is also put forward, therefore, the redundant learning after finding optimal path is avoided, and the time of learning is reduced. The validity of the algorithm is proved by simulation experiments
Keywords
learning (artificial intelligence); path planning; robots; Metropolis criterion; Q-learning algorithm; automatic termination condition; exploration region expansion strategy; optimal path finding; Blindness; Computer science; Helium; Learning; Robotics and automation; Robots; Simulated annealing; Metropolis criterion; Q-learning; exploration region expansion; exploration-exploitation;
fLanguage
English
Publisher
ieee
Conference_Titel
Intelligent Control and Automation, 2006. WCICA 2006. The Sixth World Congress on
Conference_Location
Dalian
Print_ISBN
1-4244-0332-4
Type
conf
DOI
10.1109/WCICA.2006.1713159
Filename
1713159
Link To Document