DocumentCode :
3628990
Title :
Consistency of fuzzy model-based reinforcement learning
Author :
Lucian Busoniu;Damien Ernst;Bart De Schutter;Robert Babuska
Author_Institution :
Delft Center for Systems and Control of the Delft University of Technology, The Netherlands
fYear :
2008
fDate :
6/1/2008 12:00:00 AM
Firstpage :
518
Lastpage :
524
Abstract :
Reinforcement learning (RL) is a widely used paradigm for learning control. Computing exact RL solutions is generally only possible when process states and control actions take values in a small discrete set. In practice, approximate algorithms are necessary. In this paper, we propose an approximate, model-based Q-iteration algorithm that relies on a fuzzy partition of the state space, and on a discretization of the action space. Using assumptions on the continuity of the dynamics and of the reward function, we show that the resulting algorithm is consistent, i.e., that the optimal solution is obtained asymptotically as the approximation accuracy increases. An experimental study indicates that a continuous reward function is also important for a predictable improvement in performance as the approximation accuracy increases.
Keywords :
"Approximation methods","Approximation algorithms","Magnetic cores","Convergence","Aerospace electronics","Distance measurement","Fuzzy sets"
Publisher :
ieee
Conference_Titel :
Fuzzy Systems, 2008. FUZZ-IEEE 2008. (IEEE World Congress on Computational Intelligence). IEEE International Conference on
ISSN :
1098-7584
Print_ISBN :
978-1-4244-1818-3
Type :
conf
DOI :
10.1109/FUZZY.2008.4630417
Filename :
4630417
Link To Document :
بازگشت