DocumentCode :
2955860
Title :
Learning policies for abstract state spaces
Author :
Timmer, Stephan ; Riedmiller, Martin
Author_Institution :
Neuroinformatics, Osnabrueck Univ., Germany
Volume :
4
fYear :
2005
fDate :
10-12 Oct. 2005
Firstpage :
3179
Abstract :
Applying Q-learning to multidimensional, real-valued state spaces is time-consuming in most cases. In this article, we deal with the assumption that a coarse partition of the state space is sufficient for learning good or even optimal policies. An algorithm is presented which constructs proper policies for abstract state spaces using an incremental procedure without approximating a Q-function. By combining an approach similar to dynamic programming and a search for policies, we can speed up the learning process. To provide empirical evidence, we use a cart-pole system. Experiments were conducted for a simulated environment as well as for a real plant.
Keywords :
dynamic programming; learning (artificial intelligence); state-space methods; Q-learning; abstract state space; cart-pole system; dynamic programming; incremental learning; learning policy; learning process; Cost function; Dynamic programming; Iterative algorithms; Multidimensional systems; Partitioning algorithms; State-space methods;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Systems, Man and Cybernetics, 2005 IEEE International Conference on
Print_ISBN :
0-7803-9298-1
Type :
conf
DOI :
10.1109/ICSMC.2005.1571635
Filename :
1571635
Link To Document :
بازگشت