DocumentCode :
2717289
Title :
Model-Based Reinforcement Learning in Factored-State MDPs
Author :
Strehl, Alexander L.
Author_Institution :
Dept. of Comput. Sci., Rutgers Univ.
fYear :
2007
fDate :
1-5 April 2007
Firstpage :
103
Lastpage :
110
Abstract :
We consider the problem of learning in a factored-state Markov decision process that is structured to allow a compact representation. We show that the well-known algorithm, factored Rmax, performs near-optimally on all but a number of timesteps that is polynomial in the size of the compact representation, which is often exponentially smaller than the number of states. This is equivalent to the result obtained by Kearns and Roller for their DBN-E3 algorithm, except that we´ve conducted the analysis in a more general setting. We also extend the results to a new algorithm, factored IE, that uses the interval estimation approach to exploration and can be expected to outperform factored Rmax on most domains
Keywords :
Markov processes; learning (artificial intelligence); factored-state MDPs; factored-state Markov decision process; model-based reinforcement learning; Algorithm design and analysis; Bayesian methods; Computer science; Dynamic programming; Learning; Linear approximation; Mathematical model; Performance analysis; Polynomials; State-space methods;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Approximate Dynamic Programming and Reinforcement Learning, 2007. ADPRL 2007. IEEE International Symposium on
Conference_Location :
Honolulu, HI
Print_ISBN :
1-4244-0706-0
Type :
conf
DOI :
10.1109/ADPRL.2007.368176
Filename :
4220821
Link To Document :
بازگشت