Title :
Model-based reinforcement learning for infinite-horizon approximate optimal tracking
Author :
Kamalapurkar, Rushikesh ; Andrews, Lindsey ; Walters, Patrick ; Dixon, Warren E.
Author_Institution :
Dept. of Mech. & Aerosp. Eng., Univ. of Florida, Gainesville, FL, USA
Abstract :
This paper provides an approximate online adaptive solution to the infinite-horizon optimal tracking problem for control-affine continuous-time nonlinear systems with unknown drift dynamics where model-based reinforcement learning is used to relax the persistence of excitation condition. Model-based reinforcement learning is implemented using a concurrent learning-based system identifier to simulate experience by evaluating the Bellman error over unexplored areas of the state space. Tracking of the desired trajectory and convergence of the developed policy to a neighborhood of the optimal policy is established via Lyapunov-based stability analysis.
Keywords :
Lyapunov methods; approximation theory; continuous time systems; convergence of numerical methods; learning (artificial intelligence); nonlinear control systems; stability; state-space methods; Bellman error; Lyapunov-based stability analysis; approximate online adaptive solution; concurrent learning-based system identifier; control-affine continuous-time nonlinear systems; excitation condition; infinite-horizon approximate optimal tracking; model-based reinforcement learning; optimal policy; state space; trajectory tracking; unknown drift dynamics; Artificial neural networks; Function approximation; Optimal control; Stability analysis; Steady-state; Trajectory;
Conference_Titel :
Decision and Control (CDC), 2014 IEEE 53rd Annual Conference on
Conference_Location :
Los Angeles, CA
Print_ISBN :
978-1-4799-7746-8
DOI :
10.1109/CDC.2014.7040183