DocumentCode :
337653
Title :
Stability and convergence of stochastic approximation using the ODE method
Author :
Borkar, V.S. ; Meyn, S.P.
Author_Institution :
Dept. of Comput. Sci. & Autom., Indian Inst. of Sci., Bangalore, India
Volume :
1
fYear :
1998
fDate :
1998
Firstpage :
277
Abstract :
It is shown that the stability of the stochastic approximation algorithm is implied by the asymptotic stability of the origin for an associated ODE. This in turn implies convergence of the algorithm. Several specific classes of algorithms are considered as applications. It is found that the results provide: 1) a simpler derivation of known results for reinforcement learning algorithms; 2) a proof for the first time that a class of asynchronous stochastic approximation algorithms are convergent without using any a priori assumption of stability; and 3) a proof for the first time that asynchronous adaptive critic and Q-learning algorithms are convergent for the average cost optimal control problem
Keywords :
Markov processes; approximation theory; asymptotic stability; convergence; decision theory; differential equations; learning (artificial intelligence); optimal control; stochastic processes; Markov decision process; adaptive control; asymptotic stability; convergence; optimal control; reinforcement learning; stochastic approximation; Application software; Approximation algorithms; Asymptotic stability; Automation; Computer science; Convergence; Cost function; Learning; Optimal control; Stochastic processes;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Decision and Control, 1998. Proceedings of the 37th IEEE Conference on
Conference_Location :
Tampa, FL
ISSN :
0191-2216
Print_ISBN :
0-7803-4394-8
Type :
conf
DOI :
10.1109/CDC.1998.760684
Filename :
760684
Link To Document :
بازگشت