DocumentCode
1665570
Title
Receding Horizon Cache and Extreme Learning Machine based Reinforcement Learning
Author
Zhifei Shao ; Meng Joo Er ; Guang-Bin Huang
Author_Institution
Sch. of Electr. & Electron. Eng., Nanyang Technol. Univ., Singapore, Singapore
fYear
2012
Firstpage
1591
Lastpage
1596
Abstract
Function approximators have been extensively used in Reinforcement Learning (RL) to deal with large or continuous space problems. However, batch learning Neural Networks (NN), one of the most common approximators, has been rarely applied to RL. In this paper, possible reasons for this are laid out and a solution is proposed. Specifically, a Receding Horizon Cache (RHC) structure is designed to collect training data for NN by dynamically archiving state-action pairs and actively updating their Q-values, which makes batch learning NN much easier to implement. Together with Extreme Learning Machine (ELM), a new RL with function approximation algorithm termed as RHC and ELM based RL (RHC-ELM-RL) is proposed. A mountain car task was carried out to test RHC-ELM-RL and compare its performance with other algorithms.
Keywords
approximation theory; learning (artificial intelligence); neural nets; ELM; NN; RHC structure; RL; batch learning Neural Networks; continuous space problems; extreme learning machine based reinforcement learning; function approximation algorithm; mountain car task; receding horizon cache structure; Approximation algorithms; Artificial neural networks; Educational institutions; Function approximation; Heuristic algorithms; Training; Training data;
fLanguage
English
Publisher
ieee
Conference_Titel
Control Automation Robotics & Vision (ICARCV), 2012 12th International Conference on
Conference_Location
Guangzhou
Print_ISBN
978-1-4673-1871-6
Electronic_ISBN
978-1-4673-1870-9
Type
conf
DOI
10.1109/ICARCV.2012.6485384
Filename
6485384
Link To Document