Title :
Consistency of fuzzy model-based reinforcement learning
Author :
Lucian Busoniu;Damien Ernst;Bart De Schutter;Robert Babuska
Author_Institution :
Delft Center for Systems and Control of the Delft University of Technology, The Netherlands
fDate :
6/1/2008 12:00:00 AM
Abstract :
Reinforcement learning (RL) is a widely used paradigm for learning control. Computing exact RL solutions is generally only possible when process states and control actions take values in a small discrete set. In practice, approximate algorithms are necessary. In this paper, we propose an approximate, model-based Q-iteration algorithm that relies on a fuzzy partition of the state space, and on a discretization of the action space. Using assumptions on the continuity of the dynamics and of the reward function, we show that the resulting algorithm is consistent, i.e., that the optimal solution is obtained asymptotically as the approximation accuracy increases. An experimental study indicates that a continuous reward function is also important for a predictable improvement in performance as the approximation accuracy increases.
Keywords :
"Approximation methods","Approximation algorithms","Magnetic cores","Convergence","Aerospace electronics","Distance measurement","Fuzzy sets"
Conference_Titel :
Fuzzy Systems, 2008. FUZZ-IEEE 2008. (IEEE World Congress on Computational Intelligence). IEEE International Conference on
Print_ISBN :
978-1-4244-1818-3
DOI :
10.1109/FUZZY.2008.4630417