DocumentCode :
2203026
Title :
Learning to plan probabilistically from neural networks
Author :
Sun, Ron ; Sessions, Chad
Author_Institution :
Alabama Univ., Tuscaloosa, AL, USA
Volume :
1
fYear :
1998
fDate :
4-8 May 1998
Firstpage :
1
Abstract :
This paper discusses the learning of probabilistic planning without a priori domain-specific knowledge. Different from existing reinforcement learning algorithms that generate only reactive policies and existing probabilistic planning algorithms that requires a substantial amount of a priori knowledge in order to plan, we devise a two-stage bottom-up learning-to-plan process, in which the reinforcement learning/dynamic programming is first applied, without the use of a priori domain-specific knowledge, to acquire a reactive policy and then explicit plans are extracted from the learned reactive policy. Plan extraction is based on a beam search algorithm that performs temporal projection in a restricted fashion guided by the value functions resulting from the reinforcement learning/dynamic programming. The experiments and theoretical analysis are presented
Keywords :
dynamic programming; learning (artificial intelligence); neural nets; planning (artificial intelligence); probability; search problems; beam search algorithm; bottom-up learning; dynamic programming; neural networks; plan extraction; probabilistic planning; probability; reactive policy; reinforcement learning; Artificial intelligence; Dynamic programming; Guidelines; Learning; Monitoring; Neural networks; Probability; Process planning; Sun; Welding;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks Proceedings, 1998. IEEE World Congress on Computational Intelligence. The 1998 IEEE International Joint Conference on
Conference_Location :
Anchorage, AK
ISSN :
1098-7576
Print_ISBN :
0-7803-4859-1
Type :
conf
DOI :
10.1109/IJCNN.1998.682226
Filename :
682226
Link To Document :
بازگشت