DocumentCode :
417044
Title :
A reinforcement learning accelerated by state space reduction
Author :
Senda, K. ; Mano, S. ; Fuji, S.
Author_Institution :
Kanazawa Univ., Japan
Volume :
2
fYear :
2003
fDate :
4-6 Aug. 2003
Firstpage :
1992
Abstract :
This paper discusses a method to accelerate reinforcement learning. A concept is firstly defined, i.e., the state space reduction conserving policy. An algorithm is then given, where the optimal cost-to-go and the optimal policy of the reduced space is calculated from those of the original space. Using the reduced state space, learning convergence is accelerated. Its usefulness for DP iteration and Q-learning are compared through a maze example. The convergence of the optimal cost-to-go in the original state space needs approximately N or more times as long as that in the reduced state space, where N is a ratio of the number of the original states to the reduced. The acceleration effect for Q-learning is more remarkable than that for the DP iteration. The proposal technique is also applied to a robot manipulator working for a peg-in-hole task with geometric constraints. The state space reduction can be considered as a model of the change of observation, i.e., one of cognitive actions. The obtained results explain that the change of observation is reasonable in terms of learning efficiency.
Keywords :
Markov processes; convergence of numerical methods; dynamic programming; infinite horizon; iterative methods; learning (artificial intelligence); manipulators; state-space methods; Markov processes; Q-learning; acceleration effect; convergence; dynamic programming iteration; geometric constraints; infinite horizon; optimal cost; optimal policy; peg-in-hole task; reinforcement learning; robot manipulator; state space reduction conserving policy;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
SICE 2003 Annual Conference
Conference_Location :
Fukui, Japan
Print_ISBN :
0-7803-8352-4
Type :
conf
Filename :
1324287
Link To Document :
بازگشت