DocumentCode :
677972
Title :
Deep Belief Network for Modeling Hierarchical Reinforcement Learning Policies
Author :
Djurdjevic, Predrag D. ; Huber, Marco
Author_Institution :
Dept. of Comput. Sci. & Eng., Univ. of Texas at Arlington, Arlington, TX, USA
fYear :
2013
fDate :
13-16 Oct. 2013
Firstpage :
2485
Lastpage :
2491
Abstract :
Intelligent agents over their lifetime face multiple tasks that require simultaneous modeling and control of complex, initially unknown environments, observed via incomplete and uncertain observations. In such scenarios, policy learning is subject to the curse of dimensionality, leading to scaling problems for traditional Reinforcement Learning (RL). To address this, the agent has to efficiently acquire and reuse latent knowledge. One way is through Hierarchical Reinforcement Learning (HRL), which embellishes RL with a hierarchical, model-based approach to state, reward and policy representation. This paper presents a novel learning approach for HRL based on Conditional Restricted Boltzmann Machines (CRBMs). The proposed model provides a uniform means to simultaneously learn policies and associated abstract state features, and allows learning and executing hierarchical skills within a consistent, uniform network structure. In this model, learning is performed incrementally from basic grounded features to complex abstract policies based on automatically extracted latent states and rewards.
Keywords :
Boltzmann machines; belief networks; learning (artificial intelligence); CRBM; HRL; conditional restricted Boltzmann machines; deep belief network; hierarchical reinforcement learning policy modeling; intelligent agents; Abstracts; Buildings; Computational modeling; Context; Learning (artificial intelligence); Training; Vectors;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Systems, Man, and Cybernetics (SMC), 2013 IEEE International Conference on
Conference_Location :
Manchester
Type :
conf
DOI :
10.1109/SMC.2013.424
Filename :
6722177
Link To Document :
بازگشت