DocumentCode
7313
Title
Revisiting Approximate Dynamic Programming and its Convergence
Author
Heydari, Ali
Author_Institution
Dept. of Mech. Eng., South Dakota Sch. of Mines & Technol., Rapid City, SD, USA
Volume
44
Issue
12
fYear
2014
fDate
Dec. 2014
Firstpage
2733
Lastpage
2743
Abstract
Value iteration-based approximate/adaptive dynamic programming (ADP) as an approximate solution to infinite-horizon optimal control problems with deterministic dynamics and continuous state and action spaces is investigated. The learning iterations are decomposed into an outer loop and an inner loop. A relatively simple proof for the convergence of the outer-loop iterations to the optimal solution is provided using a novel idea with some new features. It presents an analogy between the value function during the iterations and the value function of a fixed-final-time optimal control problem. The inner loop is utilized to avoid the need for solving a set of nonlinear equations or a nonlinear optimization problem numerically, at each iteration of ADP for the policy update. Sufficient conditions for the uniqueness of the solution to the policy update equation and for the convergence of the inner-loop iterations to the solution are obtained. Afterwards, the results are formed as a learning algorithm for training a neurocontroller or creating a look-up table to be used for optimal control of nonlinear systems with different initial conditions. Finally, some of the features of the investigated method are numerically analyzed.
Keywords
dynamic programming; infinite horizon; iterative methods; learning systems; neurocontrollers; nonlinear control systems; nonlinear equations; optimal control; table lookup; ADP; action spaces; adaptive dynamic programming; approximate dynamic programming; continuous state spaces; convergence; deterministic dynamics; fixed-final-time optimal control problem; infinite-horizon optimal control problems; learning iterations; look-up table; neurocontroller training; nonlinear equations; nonlinear optimization problem; nonlinear systems; outer-loop iterations; policy update equation; value function; value iteration-based approximate; Approximation methods; Convergence; Dynamic programming; Equations; Mathematical model; Optimal control; Vectors; Approximate dynamic programming; nonlinear control systems; optimal control;
fLanguage
English
Journal_Title
Cybernetics, IEEE Transactions on
Publisher
ieee
ISSN
2168-2267
Type
jour
DOI
10.1109/TCYB.2014.2314612
Filename
6815973
Link To Document