Title :
State aggregation based linear programming approach to approximate dynamic programming
Author :
Darbha, S. ; Krishnamoorthy, K. ; Pachter, M. ; Chandler, P.
Author_Institution :
Dept. of Mech. Eng., Texas A&M Univ., College Station, TX, USA
Abstract :
One often encounters the curse of dimensionality in the application of dynamic programming to determine optimal policies for controlled Markov chains. In this paper, we provide a method to construct sub-optimal policies along with a bound for the deviation of such a policy from the optimum through the use of restricted linear programming. The novelty of this approach lies in circumventing the need for a value iteration or a linear program defined on the entire state-space. Instead, the state-space is partitioned based on the reward structure and the optimal cost-to-go or value function is approximated by a constant over each partition. We associate a meta-state with each partition, where the transition probabilities between these meta-states can be derived from the original Markov chain specification. The state aggregation approach results in a significant reduction in the computational burden and lends itself to a restricted linear program defined on the aggregated state-space. Finally, the proposed method is bench marked on a perimeter surveillance stochastic control problem.
Keywords :
Markov processes; approximation theory; dynamic programming; iterative methods; linear programming; stochastic systems; controlled Markov chains; dynamic programming; optimal cost-to-go; perimeter surveillance stochastic control problem; restricted linear programming; state aggregation based linear programming; transition probabilities; value function; value iteration; Approximation methods; Delay; Equations; Linear programming; Markov processes; Silicon; Unmanned aerial vehicles;
Conference_Titel :
Decision and Control (CDC), 2010 49th IEEE Conference on
Conference_Location :
Atlanta, GA
Print_ISBN :
978-1-4244-7745-6
DOI :
10.1109/CDC.2010.5717627