DocumentCode :
3068842
Title :
Gain/variability tradeoffs in undiscounted Markov decision processes
Author :
Filar, J.A. ; Huey-Miin Lee
Author_Institution :
The Johns Hopkins University, Baltimore, Maryland
fYear :
1985
fDate :
11-13 Dec. 1985
Firstpage :
1106
Lastpage :
1112
Abstract :
We consider a finite state/action Markov Decision Process over the infinite time horizon, and with the limiting average reward criterion. However, we are interested not only in maximizing the above reward criterion but also in minimizing "the variability" of the stream of rewards. The latter notion is formalized in two alternative ways: one in terms of measuring absolute deviations from the "optimal" reward, and the other in terms of a "long-run variance" of a policy. In both cases we formulate a bi-objective optimization problem and show that efficient (i.e., "nondominated") deterministic stationary policies exist and can be computed by finite algorithms. In addition, in the former case we give an algorithm for computing a finite set of "critical efficient policies" which in a sense constitutes one complete set of reasonable responses by a decision-maker sensitive to the variability of rewards. However, the analysis of this case is intended primarily as a "sensitivity analysis" technique rather than a complete theoretical treatment of the gain/variability trade-offs.
Keywords :
Computational modeling; Costs; Random variables; Scattering; Sensitivity analysis;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Decision and Control, 1985 24th IEEE Conference on
Conference_Location :
Fort Lauderdale, FL, USA
Type :
conf
DOI :
10.1109/CDC.1985.268672
Filename :
4048472
Link To Document :
بازگشت