DocumentCode :
3071924
Title :
Reward revision for partially observed Markov decision processes
Author :
White, C.C. ; Scherer, W.T.
Author_Institution :
University of Virginia
fYear :
1985
fDate :
11-13 Dec. 1985
Firstpage :
1822
Lastpage :
1827
Abstract :
We present an algorithm for accelerating a successive approximation solution procedure for the infinite horizon, expected discounted total reward partially observed Markov decision process (POMDP). This algorithm proceeds by simultaneously constructing and solving a sequence of dynamic programs, each of which is more computationally attractive than the dynamic program associated with the original POMDP. The form of each of these dynamic programs is determined by the solution of the preceding dynamic program. A numerical study is presented that indicates the potentially significant computational value of this algorithm.
Keywords :
Acceleration; Approximation algorithms; Equations; Infinite horizon; Iterative algorithms; Process control; State-space methods; Statistics; Systems engineering and theory; Uncertainty;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Decision and Control, 1985 24th IEEE Conference on
Conference_Location :
Fort Lauderdale, FL, USA
Type :
conf
DOI :
10.1109/CDC.1985.268877
Filename :
4048636
Link To Document :
بازگشت