DocumentCode :
1798017
Title :
Approximate model-assisted Neural Fitted Q-Iteration
Author :
Lampe, Thomas ; Riedmiller, Martin
Author_Institution :
Dept. of Comput. Sci., Albert-Ludwigs-Univ., Freiburg, Germany
fYear :
2014
fDate :
6-11 July 2014
Firstpage :
2698
Lastpage :
2704
Abstract :
In this work, we propose an extension to the Neural Fitted Q-Iteration algorithm that utilizes a learned model to generate virtual trajectories which are used for updating the Q-function. Compared to standard NFQ, this combination has the potential to greatly reduce the amount of system interaction required to learn a good policy. At the same time, the approach still maintains the generalization ability of Q-learning. We provide a general formulation for approximate model-assisted fitted Q-learning, and examine the advantages of its neural implementation regarding interaction time and robustness. Its capabilities are illustrated with first results on a benchmark cart-pole regulation task, on which our method turns out to provide more general policies using much less interaction time.
Keywords :
learning (artificial intelligence); neural nets; Q-function; approximate model-assisted neural fitted Q-iteration; benchmark cart-pole regulation task; generalization ability; learned model; neural implementation; standard NFQ; virtual trajectories; Approximation algorithms; Data models; Robustness; Standards; Testing; Training; Trajectory;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks (IJCNN), 2014 International Joint Conference on
Conference_Location :
Beijing
Print_ISBN :
978-1-4799-6627-1
Type :
conf
DOI :
10.1109/IJCNN.2014.6889733
Filename :
6889733
Link To Document :
بازگشت