Title :
Comparison of reinforcement algorithms on discrete functions: learnability, time complexity, and scaling
Author :
Markey, Kevin L. ; Mozer, Michael C.
Author_Institution :
Colorado Univ., Boulder, CO, USA
Abstract :
The authors compare the performances of a variety of algorithms in a reinforcement learning paradigm, including Ar-p, Ar-i, reinforcement-comparison (plus a new variation), and backpropagation of reinforcement gradient through a forward model. The task domain is discrete multioutput functions. Performance is measured in terms of learnability, training time, and scaling. Ar-p outperforms all others and scales well relative to supervised backpropagation. An ergodic variant of reinforcement-comparison approaches Ar-p performance. For the tasks studied, total training time (including model and controller) for the forward model algorithm is 1 to 2 orders of magnitude more costly than for Ar-p, and the controller´s success is sensitive to forward model accuracy. Distortions of the reinforcement gradient predicted by an inaccurate forward model cause the controller´s failures
Keywords :
backpropagation; computational complexity; learning (artificial intelligence); neural nets; Ar-i; Ar-p; backpropagation; discrete functions; ergodic variant; forward model; forward model algorithm; learnability; reinforcement algorithms; reinforcement gradient; scaling; task domain; time complexity; training time; Algorithm design and analysis; Backpropagation algorithms; Cognitive science; Computer science; Control systems; Learning systems; Predictive models; Signal design; Stochastic processes; Time measurement;
Conference_Titel :
Neural Networks, 1992. IJCNN., International Joint Conference on
Conference_Location :
Baltimore, MD
Print_ISBN :
0-7803-0559-0
DOI :
10.1109/IJCNN.1992.287080