Author :
Liang, Tian ; Sun, Jigui ; Yin, Minghao
Abstract :
Real-time dynamic programming (RTDP) is an outstanding real-time algorithm for solving non-deterministic planning problems with full observability. RTDP has two key advantages comparing with other DP algorithms: first, it obtain an optimal policy without computing the whole space, second, it has a good anytime behavior. However, RTDP\´s convergence is slow. In this paper, we introduce RTDP(k), an algorithm based on RTDP with a similar structure. RTDP(k) improves the convergence as well as holds real-time algorithm properties. RTDP(k) updates k extended states per iteration following a "bounded propagation strategy". In Markov decision processes (MDPs), especially in stochastic shortest-path problems (SSPs), we have proved two points: first, every RTDP(k) trial terminates in a finite number of steps ,second, RTDP(k) eventually converges to an optimal policy. From a practical point of view, We show that RTDP(k) produces better solutions in the first trial and converges faster than RTDP on benchmarks for real-time search.
Keywords :
Markov processes; decision theory; dynamic programming; graph theory; iterative methods; planning; search problems; stochastic programming; Markov decision pocesses; bounded propagation strategy; iteration; nondeterministic planning problems; optimal policy; ral-time dynamic programming; real-time search; stochastic shortest-path problems; Convergence; Cost function; Dynamic programming; Educational institutions; Labeling; Observability; Probability distribution; State-space methods; Stochastic processes; Sun;