DocumentCode :
1199172
Title :
Semi-Markov decision problems and performance sensitivity analysis
Author :
Cao, Xi-Ren
Author_Institution :
Hong Kong Univ. of Sci. & Technol., China
Volume :
48
Issue :
5
fYear :
2003
fDate :
5/1/2003 12:00:00 AM
Firstpage :
758
Lastpage :
769
Abstract :
Recent research indicates that Markov decision processes (MDPs) can be viewed from a sensitivity point of view; and the perturbation analysis (PA), MDPs, and reinforcement learning (RL) are three closely related areas in optimization of discrete-event dynamic systems that can be modeled as Markov processes. The goal of this paper is two-fold. First, we develop the PA theory for semi-Markov processes (SMPs); and then we extend the aforementioned results about the relation among PA, MDP, and RL to SMPs. In particular, we show that performance sensitivity formulas and policy iteration algorithms of semi-Markov decision processes can be derived based on the performance potential and realization matrix. Both the long-run average and discounted-cost problems are considered. This approach provides a unified framework for both problems, and the long-run average problem corresponds to the discounted factor being zero. The results indicate that performance sensitivities and optimization depend only on first-order statistics. Single sample path-based implementations are discussed.
Keywords :
Lyapunov methods; Markov processes; discrete event systems; iterative methods; optimisation; perturbation techniques; sensitivity analysis; Lyapunov equations; Markov decision processes; Poisson equations; discounted Poisson equations; discrete-event dynamic systems; iteration algorithms; perturbation analysis; policy iteration; reinforcement learning; sensitivity analysis; Learning; Markov processes; Performance analysis; Poisson equations; Queueing analysis; Sensitivity analysis; State estimation; Statistics; Stochastic processes; User-generated content;
fLanguage :
English
Journal_Title :
Automatic Control, IEEE Transactions on
Publisher :
ieee
ISSN :
0018-9286
Type :
jour
DOI :
10.1109/TAC.2003.811252
Filename :
1198597
Link To Document :
بازگشت