Title :
An Improved Q-learning Algorithm Based on Exploration Region Expansion Strategy
Author :
Gao, Qingji ; Hong, Bingrong ; He, Zhendong ; Liu, Jie ; Niu, Guochen
Author_Institution :
Dept. of Comput. Sci. & Technol., Harbin Inst. of Technol.
Abstract :
In order to find a good solution to one of the key problems in Q-learning algorithm - keeping the balance between exploration and exploitation, an improved Q-learning algorithm based on exploration region expansion strategy is proposed on the base of Metropolis criterion-based Q-learning. With this strategy, the exploration blindness in the entire environment is eliminated, and the learning efficiency is increased. Meanwhile, other feasible path is sought where agent encounters obstacles, which makes the implementation of the algorithm on real robot easy. An automatic termination condition is also put forward, therefore, the redundant learning after finding optimal path is avoided, and the time of learning is reduced. The validity of the algorithm is proved by simulation experiments
Keywords :
learning (artificial intelligence); path planning; robots; Metropolis criterion; Q-learning algorithm; automatic termination condition; exploration region expansion strategy; optimal path finding; Blindness; Computer science; Helium; Learning; Robotics and automation; Robots; Simulated annealing; Metropolis criterion; Q-learning; exploration region expansion; exploration-exploitation;
Conference_Titel :
Intelligent Control and Automation, 2006. WCICA 2006. The Sixth World Congress on
Conference_Location :
Dalian
Print_ISBN :
1-4244-0332-4
DOI :
10.1109/WCICA.2006.1713159