• DocumentCode
    1752965
  • Title

    An Improved Q-learning Algorithm Based on Exploration Region Expansion Strategy

  • Author

    Gao, Qingji ; Hong, Bingrong ; He, Zhendong ; Liu, Jie ; Niu, Guochen

  • Author_Institution
    Dept. of Comput. Sci. & Technol., Harbin Inst. of Technol.
  • Volume
    1
  • fYear
    0
  • fDate
    0-0 0
  • Firstpage
    4167
  • Lastpage
    4170
  • Abstract
    In order to find a good solution to one of the key problems in Q-learning algorithm - keeping the balance between exploration and exploitation, an improved Q-learning algorithm based on exploration region expansion strategy is proposed on the base of Metropolis criterion-based Q-learning. With this strategy, the exploration blindness in the entire environment is eliminated, and the learning efficiency is increased. Meanwhile, other feasible path is sought where agent encounters obstacles, which makes the implementation of the algorithm on real robot easy. An automatic termination condition is also put forward, therefore, the redundant learning after finding optimal path is avoided, and the time of learning is reduced. The validity of the algorithm is proved by simulation experiments
  • Keywords
    learning (artificial intelligence); path planning; robots; Metropolis criterion; Q-learning algorithm; automatic termination condition; exploration region expansion strategy; optimal path finding; Blindness; Computer science; Helium; Learning; Robotics and automation; Robots; Simulated annealing; Metropolis criterion; Q-learning; exploration region expansion; exploration-exploitation;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Intelligent Control and Automation, 2006. WCICA 2006. The Sixth World Congress on
  • Conference_Location
    Dalian
  • Print_ISBN
    1-4244-0332-4
  • Type

    conf

  • DOI
    10.1109/WCICA.2006.1713159
  • Filename
    1713159