DocumentCode :
3572792
Title :
Data-driven optimal control with reduced output measurements
Author :
Modares, Hamidreza ; Gan Oon Peen ; Lemei Zhu ; Lewis, Frank L. ; Baozeng Yue
Author_Institution :
Arlington Res. Inst., Univ. of Texas, Arlington, TX, USA
fYear :
2014
Firstpage :
1775
Lastpage :
1780
Abstract :
This paper uses the integral reinforcement learning (IRL) technique to develop an online learning algorithm for finding suboptimal static output-feedback controllers for partially-unknown continuous-time (CT) linear systems. To our knowledge, this is the first static output-feedback control design method based on reinforcement learning for CT systems. In the proposed method, an online policy iteration (PI) algorithm is developed which uses the integral reinforcement knowledge for learning a suboptimal static output-feedback solution without requiring the drift knowledge of the system dynamics. Specifically, in the policy evaluation step of the PI algorithm, an IRL Bellman equation is used to evaluate an output-feedback policy, and in the policy improvement step of the PI algorithm the output-feedback gain is updated using the information given by the evaluated policy. An adaptive observer is used to provide the knowledge of the full states for the IRL Bellman equation during learning. However, the observer is not needed after the learning process is finished. The convergence of the proposed algorithm to a suboptimal output-feedback solution and the performance of the proposed method are verified through simulations.
Keywords :
continuous time systems; dynamic programming; feedback; iterative methods; learning (artificial intelligence); linear systems; observers; optimal control; CT linear systems; IRL Bellman equation; PI algorithm; adaptive observer; continuous-time linear systems; data-driven optimal control; integral reinforcement learning technique; online learning algorithm; online policy iteration algorithm; output-feedback gain; output-feedback policy; policy evaluation step; policy improvement step; reduced output measurements; suboptimal static output-feedback controllers; Equations; Heuristic algorithms; Learning (artificial intelligence); Mathematical model; Observers; Output feedback; Integral reinforcement learning; Optimal control; Output-feedback control;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Intelligent Control and Automation (WCICA), 2014 11th World Congress on
Type :
conf
DOI :
10.1109/WCICA.2014.7052989
Filename :
7052989
Link To Document :
بازگشت