Title :
Approximate policy improvement for continuous action set-point regulation problems
Author :
Esogbue, A.O. ; Hearnes, W.E., II
Author_Institution :
Sch. of Ind. & Syst. Eng., Georgia Inst. of Technol., Atlanta, GA, USA
Abstract :
Model-free control methods based on classical dynamic programming approximation algorithms are a fertile area of research. These methods are generally for discrete state and action spaces. This research proposes an algorithm for extending approximate policy iteration to continuous action spaces by combining the derivative-free line search methods of nonlinear optimization with policy improvement based on Q-learning using a discrete subset of reference actions. The properties of the proposed algorithm are discussed and two example problems illustrate its applicability. Q-learning algorithms are used to search a continuous action space. The method reduces the computational effort required in many problems
Keywords :
dynamic programming; optimal control; search problems; Q-learning; approximate policy improvement; classical dynamic programming approximation algorithms; continuous action set-point regulation problems; derivative-free line search methods; model-free control methods; nonlinear optimization; Control systems; Dynamic programming; Intelligent control; Intelligent systems; Laboratories; Learning; Power system dynamics; Power system modeling; Process control; Space technology;
Conference_Titel :
Fuzzy Systems Proceedings, 1998. IEEE World Congress on Computational Intelligence., The 1998 IEEE International Conference on
Conference_Location :
Anchorage, AK
Print_ISBN :
0-7803-4863-X
DOI :
10.1109/FUZZY.1998.686375