Title : 
Efficient experience reuse in non-Markovian environments
         
        
            Author : 
Dung, Le Tien ; Komeda, Takashi ; Takagi, Motoki
         
        
            Author_Institution : 
Grad. Sch. of Eng., Shibaura Inst. of Technol., Tokyo
         
        
        
        
        
        
            Abstract : 
Learning time is always a critical issue in Reinforcement Learning, especially when Recurrent Neural Networks are used to predict Q values in non-Markovian environments. Experience reuse has been received much attention due to its ability to reduce learning time. In this paper, we propose a new method to efficiently reuse experience. Our method generates new episodes from recorded episodes using an action-pair merger. Recorded episodes and new episodes are replayed after each learning epoch. We compare our method with standard online learning, and learning using experience replay in a vision based robot problem. The results show the potential of this approach.
         
        
            Keywords : 
learning (artificial intelligence); recurrent neural nets; action-pair merger; learning time; nonMarkovian environments; online learning; recurrent neural networks; reinforcement learning; vision-based robot problem; Corporate acquisitions; Electronic mail; Large-scale systems; Learning; Neural networks; Recurrent neural networks; Robot vision systems; State-space methods; Stochastic processes; Systems engineering and theory; Recurrent Neural Networks; Reinforcement Learning;
         
        
        
        
            Conference_Titel : 
SICE Annual Conference, 2008
         
        
            Conference_Location : 
Tokyo
         
        
            Print_ISBN : 
978-4-907764-30-2
         
        
            Electronic_ISBN : 
978-4-907764-29-6
         
        
        
            DOI : 
10.1109/SICE.2008.4655239