DocumentCode :
493369
Title :
Adaptive computation of optimal nonrandomized policies in constrained average-reward MDPs
Author :
Feinberg, Eugene A.
Author_Institution :
Dept. of Appl. Math. & Stat., Stony Brook Univ., Stony Brook, NY
fYear :
2009
fDate :
March 30 2009-April 2 2009
Firstpage :
96
Lastpage :
100
Abstract :
This paper deals with computation of optimal nonrandomized nonstationary policies and mixed stationary policies for average-reward Markov decision processes with multiple criteria and constraints. We consider problems with finite state and action sets satisfying the unichain condition. The described procedure for computing optimal nonrandomized policies can also be used for adaptive control problems.
Keywords :
Markov processes; action sets; adaptive computation; adaptive control problems; average-reward Markov decision processes; constrained average-reward MDP; finite state; mixed stationary policies; optimal nonrandomized nonstationary policies; unichain condition; Adaptive control; Constraint theory; Frequency; Linear programming; Mathematics; Probability distribution; State-space methods; Statistics; Vectors;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Adaptive Dynamic Programming and Reinforcement Learning, 2009. ADPRL '09. IEEE Symposium on
Conference_Location :
Nashville, TN
Print_ISBN :
978-1-4244-2761-1
Type :
conf
DOI :
10.1109/ADPRL.2009.4927531
Filename :
4927531
Link To Document :
بازگشت