DocumentCode :
2949997
Title :
S-learning: a reinforcement learning method without parameter tuning
Author :
Chen, Hown-Wen ; Soo, Von-Wun
Author_Institution :
Dept. of Comput. Sci., Nat. Tsing Hua Univ., Hsinchu, Taiwan
Volume :
1
fYear :
1993
fDate :
25-29 Oct. 1993
Firstpage :
557
Abstract :
Discussed issues raised in two reinforcement learning algorithms, adaptive heuristic critic (AHC) and Q-learning. Aspects considered are convergence, parameter tuning, over-training, and computational and storage efficiency. Two new reinforcement learning mechanisms are proposed: S-learning (step-learning) and S&AHC learning. Particularly, the representation of the final cost map formed by S-learning series can be explicitly interpreted as the number of minimum movements to the goal state from each individual state. In addition, an adaptive (incremental) S-learning was proposed which incorporated S-learning and the technique of incremental learning to facilitate the practical implementation of neural reinforcement learning. All of S-learning series showed promising performances in exploring Sutton´s task (1991) of navigating in a maze.
Keywords :
computational complexity; heuristic programming; learning (artificial intelligence); AHC; Q-learning; S&AHC learning; S-learning; adaptive heuristic critic; computational efficiency; convergence; neural reinforcement learning; over-training; parameter tuning; reinforcement learning method; step-learning; storage efficiency; Biological control systems; Biological system modeling; Computer science; Convergence; Costs; Delay; Game theory; Learning; Navigation; State estimation;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 1993. IJCNN '93-Nagoya. Proceedings of 1993 International Joint Conference on
Print_ISBN :
0-7803-1421-2
Type :
conf
DOI :
10.1109/IJCNN.1993.713976
Filename :
713976
Link To Document :
بازگشت