DocumentCode :
3043699
Title :
Model-Based Indirect Learning Method Based on Dyna-Q Architecture
Author :
Kao-Shing Hwang ; Wei-Cheng Jiang ; Yu-Jen Chen ; Wei-Han Wang
Author_Institution :
Nat. Sun Yat-sen Univ., Kaohsiung, Taiwan
fYear :
2013
fDate :
13-16 Oct. 2013
Firstpage :
2540
Lastpage :
2544
Abstract :
In this paper, a model learning method based on tree structures is present to achieve the sample efficiency in stochastic environment. The proposed method is composed of Q-Learning algorithm to form a Dyna agent that can used to speed up learning. The Q-Learning is used to learn the policy, and the proposed method is for model learning. The model builds the environment model and simulates the virtual experience. The virtual experience can decrease the interaction between the agent and the environment and make the agent perform value iterations quickly. Thus, the proposed agent has additional experience for updating the policy. The simulation task, a mobile robot in a maze, is introduced to compare the methods, Q-Learning, Dyna-Q and the proposed method. The result of simulation confirms the proposed method that can achieve the goal of sample efficiency.
Keywords :
decision trees; iterative methods; learning (artificial intelligence); mobile robots; multi-agent systems; stochastic processes; Dyna agent; Dyna-Q architecture; Q-Learning; Q-Learning algorithm; mobile robot; model-based indirect learning method; stochastic environment; tree structures; virtual experience; Learning (artificial intelligence); Learning systems; Markov processes; Planning; Support vector machine classification; Training; Vectors; Dyna-Q; decision tree; model learning;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Systems, Man, and Cybernetics (SMC), 2013 IEEE International Conference on
Conference_Location :
Manchester
Type :
conf
DOI :
10.1109/SMC.2013.433
Filename :
6722186
Link To Document :
بازگشت