DocumentCode :
2919022
Title :
A comparison of supervised and reinforcement learning methods on a reinforcement learning task
Author :
Gullapalli, VijayKumar
Author_Institution :
Dept. of Comput. & Inf. Sci., Massachusetts Univ., Amherst, MA, USA
fYear :
1991
fDate :
13-15 Aug 1991
Firstpage :
394
Lastpage :
399
Abstract :
The forward modeling approach of M.I. Jordan and J.E. Rumelhart (1990) has been shown to be applicable when supervised learning methods are to be used for solving reinforcement learning tasks. Because such tasks are natural candidates for the application of reinforcement learning methods, there is a need to evaluate the relative merits of these two learning methods on reinforcement learning tasks. The author presents one such comparison on a task involving learning to control an unstable, nonminimum phase, dynamic system. The comparison shows that the reinforcement learning method used performs better than the supervised learning method. An examination of the learning behavior of the two methods indicates that the differences in performance can be attributed to the underlying mechanics of the two learning methods, which provides grounds for believing that similar performance differences can be expected on other reinforcement learning tasks as well
Keywords :
control system analysis; learning systems; cart pole learning; forward modeling; nonminimum phase dynamic systems; pole balancing; reinforcement learning; supervised learning; Application software; Information science; Learning systems; Supervised learning; Unsupervised learning;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Intelligent Control, 1991., Proceedings of the 1991 IEEE International Symposium on
Conference_Location :
Arlington, VA
ISSN :
2158-9860
Print_ISBN :
0-7803-0106-4
Type :
conf
DOI :
10.1109/ISIC.1991.187390
Filename :
187390
Link To Document :
بازگشت