Title :
Optimal tracking control for linear discrete-time systems using reinforcement learning
Author :
Kiumarsi-Khomartash, Bahare ; Lewis, Frank L. ; Naghibi-Sistani, Mohammad-Bagher ; Karimpour, Ali
Author_Institution :
Dept. of Electr. Eng., Ferdowsi Univ. of Mashhad, Mashhad, Iran
Abstract :
This paper presents an online solution to the infinite-horizon linear quadratic tracker (LQT) using reinforcement learning. It is first assumed that the value function for the LQT is quadratic in terms of the reference trajectory and the state of the system. Then, using the quadratic form of the value function, an augmented algebraic Riccati equation (ARE) is derived to solve the LQT. Using this formulation, both feedback and feedforward parts of the optimal control solution are obtained simultaneously by solving the augmented ARE. To find the solution to the augmented ARE online, policy iteration as a class of reinforcement learning algorithms, is employed. This algorithm is implemented on an actor-critic structure by using two neural networks and it does not need the knowledge of the drift system dynamics or the command generator dynamics. A simulation example shows that the proposed algorithm works for a system with partially unknown dynamics.
Keywords :
Riccati equations; discrete time systems; iterative methods; learning (artificial intelligence); linear quadratic control; linear systems; neural nets; ARE; LQT; actor-critic structure; augmented algebraic Riccati equation; command generator dynamics; drift system dynamics; infinite-horizon linear quadratic tracker; linear discrete-time systems; neural networks; optimal tracking control; policy iteration; reinforcement learning; Approximation methods; Artificial neural networks; Electronic mail; Facsimile; Generators; Riccati equations; Vectors; algebraic Riccati equation; linear quadratic tracker; policy iteration; reinforcement learning;
Conference_Titel :
Decision and Control (CDC), 2013 IEEE 52nd Annual Conference on
Conference_Location :
Firenze
Print_ISBN :
978-1-4673-5714-2
DOI :
10.1109/CDC.2013.6760476