Title :
Reinforcement learning with reference tracking control in continuous state spaces
Author :
Hall, Joseph ; Rasmussen, Carl Edward ; Maciejowski, Jan
Author_Institution :
Dept. of Eng., Cambridge Univ., Cambridge, UK
Abstract :
The contribution described in this paper is an algorithm for learning nonlinear, reference tracking, control policies given no prior knowledge of the dynamical system and limited interaction with the system through the learning process. Concepts from the field of reinforcement learning, Bayesian statistics and classical control have been brought together in the formulation of this algorithm which can be viewed as a form of indirect self tuning regulator. On the task of reference tracking using a simulated inverted pendulum it was shown to yield generally improved performance on the best controller derived from the standard linear quadratic method using only 30 s of total interaction with the system. Finally, the algorithm was shown to work on the simulated double pendulum proving its ability to solve nontrivial control tasks.
Keywords :
Bayes methods; adaptive control; learning (artificial intelligence); linear quadratic control; nonlinear control systems; pendulums; self-adjusting systems; state-space methods; Bayesian statistics; continuous state spaces; dynamical system; indirect self tuning regulator; nonlinear control policies; reference tracking control; reinforcement learning process; simulated double pendulum; simulated inverted pendulum; standard linear quadratic method; Data models; Gaussian processes; Heuristic algorithms; Learning; Trajectory; Uncertainty; Vectors;
Conference_Titel :
Decision and Control and European Control Conference (CDC-ECC), 2011 50th IEEE Conference on
Conference_Location :
Orlando, FL
Print_ISBN :
978-1-61284-800-6
Electronic_ISBN :
0743-1546
DOI :
10.1109/CDC.2011.6161108