Title :
A study on abstract policy for acceleration of reinforcement learning
Author :
Mohd Faudzi, Ahmad Athif ; Takano, Hirotaka ; Murata, Junichi
Author_Institution :
Dept. of Electr. & Electron. Eng., Kyushu Univ., Fukuoka, Japan
Abstract :
Reinforcement learning (RL) is well known as one of the methods that can be applied to unknown problems. However, because optimization at every state requires trial-and-error, the learning time becomes large when environment has many states. If there exist solutions to similar problems and they are used during the exploration, some of trial-and-error can be spared and the learning can take a shorter time. In this paper, the authors propose to reuse an abstract policy, a representative of a solution constructed by learning vector quantization (LVQ) algorithm, to improve initial performance of an RL learner in a similar but different problem. Furthermore, it is investigated whether or not the policy can adapt to a new environment while preserving its performance in the old environments. Simulations show good result in terms of the learning acceleration and the adaptation of abstract policy.
Keywords :
learning (artificial intelligence); vector quantisation; LVQ algorithm; RL learner; abstract policy; learning vector quantization algorithm; reinforcement learning acceleration; Abstracts; Acceleration; Educational institutions; Learning (artificial intelligence); Learning systems; Vector quantization; Vectors; Q-learning; abstraction; learning vector quantization; prior information;
Conference_Titel :
SICE Annual Conference (SICE), 2014 Proceedings of the
Conference_Location :
Sapporo
DOI :
10.1109/SICE.2014.6935300