Title :
A reinforcement learning approach involving a shortest path finding algorithm
Author :
Kwon, Woo Young ; Lee, Sanghoon ; Suh, II Hong
Author_Institution :
Graduate Sch. of Inf. & Commun., Hanyang Univ., Seoul, South Korea
Abstract :
Reinforcement learning (RL) has been widely used as a learning mechanism of an artificial life system. However, RL usually suffers from slow convergence to the optimum state-action sequence or a sequence of stimulus-response (SR) behaviors, and may not correctly work in non-Markov processes. In this paper, first, to cope with slow-convergence problem, if some state-action pairs are considered as disturbance for optimum sequence, then they are to be eliminated in long-term memory (LTM), where such disturbances are found by a shortest path-finding algorithm. This process is shown to let the system get an enhanced learning speed. Second, to partly solve non-Markov problem, if a stimulus is frequently met in a searching-process, then the stimulus is classified as a sequential percept for a non-Markov hidden state. And thus, a correct behavior for a non-Markov hidden state can be learned as in a Markov environment. To show the validity of our proposed learning technologies, several simulation results are illustrated.
Keywords :
Markov processes; artificial life; convergence; learning (artificial intelligence); path planning; artificial life system; enhanced learning speed; long-term memory; nonMarkov problem; reinforcement learning; shortest path finding algorithm; state-action sequence; stimulus-response; Animals; Convergence; Delay; History; Learning systems; Monte Carlo methods; Organisms; Signal processing; Strontium; Virtual environment;
Conference_Titel :
Intelligent Robots and Systems, 2003. (IROS 2003). Proceedings. 2003 IEEE/RSJ International Conference on
Print_ISBN :
0-7803-7860-1
DOI :
10.1109/IROS.2003.1250668