DocumentCode :
1513284
Title :
On State Aggregation to Approximate Complex Value Functions in Large-Scale Markov Decision Processes
Author :
Jia, Qing-Shan
Author_Institution :
Dept. of Autom., Tsinghua Univ., Beijing, China
Volume :
56
Issue :
2
fYear :
2011
Firstpage :
333
Lastpage :
344
Abstract :
Many small electronic devices such as cell phones and wireless sensors have restrictive memory space, computing power, and battery. The pervasive applications of these devices in industry, military, and our daily lives require simple policies that are easy to implement, and can be executed in a reasonably short time. The Markov decision process (MDP) provides a general framework for these policy optimization problems with complexity constraint. In many cases, we can use a powerful computer to find the optimal (or a good) policy and the value function first, and then approximate by a simple one. The approximation usually depends on heuristics or experiences because the relationship between the complexity of a function and the approximation error is not clear in general. In this paper we assume the optimal value function is known (or a reasonably good estimate is available) and consider how to approximate a complex value function. Due to the broad application of state aggregation in the large-scale MDP, we focus on piecewise constant approximate value functions and use the number of aggregated states to measure the complexity of a value function. We quantify how the complexity of a value function affects the approximation error. When the optimal value function is known for sure we develop an algorithm that finds the best simple state aggregation within polynomial time. When we have estimates of the optimal value function, we apply ordinal optimization to find good simple state aggregations with high probability. The algorithms are demonstrated on a node activation policy optimization problem in wireless sensor network. We hope this work can shed some insight on how to find simple policies with good performances.
Keywords :
Markov processes; computational complexity; function approximation; optimisation; piecewise constant techniques; wireless sensor networks; complex value function approximation; large-scale Markov decision process; node activation policy optimization problem; optimal value function; piecewise constant approximate value functions; polynomial time; state aggregation; wireless sensor network; Descriptive complexity; Markov decision processes (MDPs); discrete event dynamic systems; ordinal optimization; state aggregation;
fLanguage :
English
Journal_Title :
Automatic Control, IEEE Transactions on
Publisher :
ieee
ISSN :
0018-9286
Type :
jour
DOI :
10.1109/TAC.2010.2052697
Filename :
5483078
Link To Document :
بازگشت