DocumentCode :
3572785
Title :
A unified approach for semi-Markov decision processes with discounted and average reward criteria
Author :
Yanjie Li ; Huijing Wang ; Haoyao Chen
Author_Institution :
Shenzhen Grad. Sch., Harbin Inst. of Technol., Shenzhen, China
fYear :
2014
Firstpage :
1741
Lastpage :
1744
Abstract :
On the basis of the sensitivity-based optimization, we develop a unified optimization approach for semi-Markov decision processes (SMDPs) with infinite horizon discounted and average reward criteria. We show that the sensitivity formula under average reward criteria is a limitation case of discounted reward criteria. On the basis of the performance sensitivity formulas, we provide a unified formulation for the policy iteration algorithms of semi-Markov decision processes with discounted and average reward criteria.
Keywords :
Markov processes; infinite horizon; iterative methods; optimisation; sensitivity; SMDP; average reward criteria; discounted reward criteria; infinite horizon discounted; performance sensitivity formula; policy iteration algorithm; semi-Markov decision process; sensitivity-based optimization; unified optimization approach; Algorithm design and analysis; Educational institutions; Equations; Markov processes; Optimization; Sensitivity; Vectors; SMDPs; performance difference; policy iteration;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Intelligent Control and Automation (WCICA), 2014 11th World Congress on
Type :
conf
DOI :
10.1109/WCICA.2014.7052983
Filename :
7052983
Link To Document :
بازگشت