Title :
Efficient experience reuse in non-Markovian environments
Author :
Dung, Le Tien ; Komeda, Takashi ; Takagi, Motoki
Author_Institution :
Grad. Sch. of Eng., Shibaura Inst. of Technol., Tokyo
Abstract :
Learning time is always a critical issue in Reinforcement Learning, especially when Recurrent Neural Networks are used to predict Q values in non-Markovian environments. Experience reuse has been received much attention due to its ability to reduce learning time. In this paper, we propose a new method to efficiently reuse experience. Our method generates new episodes from recorded episodes using an action-pair merger. Recorded episodes and new episodes are replayed after each learning epoch. We compare our method with standard online learning, and learning using experience replay in a vision based robot problem. The results show the potential of this approach.
Keywords :
learning (artificial intelligence); recurrent neural nets; action-pair merger; learning time; nonMarkovian environments; online learning; recurrent neural networks; reinforcement learning; vision-based robot problem; Corporate acquisitions; Electronic mail; Large-scale systems; Learning; Neural networks; Recurrent neural networks; Robot vision systems; State-space methods; Stochastic processes; Systems engineering and theory; Recurrent Neural Networks; Reinforcement Learning;
Conference_Titel :
SICE Annual Conference, 2008
Conference_Location :
Tokyo
Print_ISBN :
978-4-907764-30-2
Electronic_ISBN :
978-4-907764-29-6
DOI :
10.1109/SICE.2008.4655239