Title :
Evolutionary adaptive-critic methods for reinforcement learning
Author :
Xu, Xin ; He, Han-gen ; Hu, Dewen
Author_Institution :
Dept. of Autom. Control, Nat. Univ. of Defense Technol., Changsha, China
fDate :
6/24/1905 12:00:00 AM
Abstract :
In this paper, a novel hybrid learning method is proposed for reinforcement learning problems with continuous state and action spaces. The reinforcement learning problems are modeled as Markov decision processes (MDPs) and the hybrid learning method combines evolutionary algorithms with gradient-based adaptive heuristic critic (AHC) algorithms to approximate the optimal policy of MDPs. The suggested method takes the advantages of evolutionary learning and gradient-based reinforcement learning to solve reinforcement learning problems. Simulation results on the learning control of an acrobot illustrate the efficiency of the presented method
Keywords :
Markov processes; decision theory; evolutionary computation; heuristic programming; learning (artificial intelligence); robots; Markov decision processes; acrobot; action spaces; continuous state spaces; evolutionary adaptive-critic methods; evolutionary algorithms; evolutionary learning; gradient-based adaptive heuristic critic algorithms; hybrid learning method; optimal policy; reinforcement learning; robots; simulation; Dynamic programming; Evolutionary computation; Helium; Heuristic algorithms; Intelligent agent; Learning systems; Optimal control; Optimization methods; Space technology; Supervised learning;
Conference_Titel :
Evolutionary Computation, 2002. CEC '02. Proceedings of the 2002 Congress on
Conference_Location :
Honolulu, HI
Print_ISBN :
0-7803-7282-4
DOI :
10.1109/CEC.2002.1004434