DocumentCode :
1604679
Title :
Learning of Deterministic Exploration and Temporal Abstraction in Reinforcement Learning
Author :
Shibata, Katsunari
Author_Institution :
Dept. of Electr. & Electron. Eng., Oita Univ.
fYear :
2006
Firstpage :
4569
Lastpage :
4574
Abstract :
Temporal abstraction and exploration are both very important factors to determine the performance in reinforcement learning. The author has proposed to focus on the deterministic exploration behavior that is obtained through reinforcement learning. In this paper, a novel idea that deterministic exploration behavior can be considered as temporally abstract actions or macro actions was introduced. It was actually shown in some simulations that the deterministic exploration behavior obtained through the learning of a task accelerates the learning of another similar task without any definition of abstract actions. A recurrent neural network was used for the learning, but the knowledge obtained through the first learning was used effectively in the second learning without being destroyed completely even though it did not work in a more difficult task. Furthermore, when the agent was returned to the first task, the learning was still faster than the learning from scratch. An interesting phenomenon was observed in the simulation that context-based exploration behavior was acquired through the learning of a task that did not require such behavior
Keywords :
learning (artificial intelligence); recurrent neural nets; context-based exploration behavior; deterministic exploration; recurrent neural network; reinforcement learning; temporal abstraction; Acceleration; Artificial intelligence; Context modeling; Intelligent actuators; Intelligent robots; Learning systems; Orbital robotics; Recurrent neural networks; State-space methods; Stochastic processes; Deterministic Exploration; Recurrent Neural Network; Reinforcement Learning; Temporal Abstraction;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
SICE-ICASE, 2006. International Joint Conference
Conference_Location :
Busan
Print_ISBN :
89-950038-4-7
Electronic_ISBN :
89-950038-5-5
Type :
conf
DOI :
10.1109/SICE.2006.315090
Filename :
4108483
Link To Document :
بازگشت