Title :
Model-Based Reinforcement Learning in Factored-State MDPs
Author :
Strehl, Alexander L.
Author_Institution :
Dept. of Comput. Sci., Rutgers Univ.
Abstract :
We consider the problem of learning in a factored-state Markov decision process that is structured to allow a compact representation. We show that the well-known algorithm, factored Rmax, performs near-optimally on all but a number of timesteps that is polynomial in the size of the compact representation, which is often exponentially smaller than the number of states. This is equivalent to the result obtained by Kearns and Roller for their DBN-E3 algorithm, except that we´ve conducted the analysis in a more general setting. We also extend the results to a new algorithm, factored IE, that uses the interval estimation approach to exploration and can be expected to outperform factored Rmax on most domains
Keywords :
Markov processes; learning (artificial intelligence); factored-state MDPs; factored-state Markov decision process; model-based reinforcement learning; Algorithm design and analysis; Bayesian methods; Computer science; Dynamic programming; Learning; Linear approximation; Mathematical model; Performance analysis; Polynomials; State-space methods;
Conference_Titel :
Approximate Dynamic Programming and Reinforcement Learning, 2007. ADPRL 2007. IEEE International Symposium on
Conference_Location :
Honolulu, HI
Print_ISBN :
1-4244-0706-0
DOI :
10.1109/ADPRL.2007.368176