DocumentCode :
395547
Title :
Anticipative reinforcement learning
Author :
Maire, Frederic
Author_Institution :
Sch. of Comput. Sci. & Software Eng., Queensland Univ. of Technol., Brisbane, Qld., Australia
Volume :
3
fYear :
2002
fDate :
18-22 Nov. 2002
Firstpage :
1428
Abstract :
This paper introduces anticipative reinforcement learning (ARL), a method that addresses the problem of the breakdown of value based algorithms for problems with small time steps and continuous action and state spaces when the algorithms are implemented with neural networks. In ARL, an agent is made of three components; the actor, the critic and the model (the model is as in Dyna but we use it differently). The main originality of ARL lies in the action selection process; the agent builds a set of candidate actions that includes the action recommended by the actor plus some random actions. Once the set of candidate actions is built, the candidate actions are ranked by considering what would happen if these actions were taken and followed by a sequence of actions using only the current policy (anticipation using iteratively the model with a finite look-ahead). We demonstrate the benefits of looking ahead with experiments on a Khepera robot.
Keywords :
function approximation; generalisation (artificial intelligence); learning (artificial intelligence); mobile robots; neural nets; state-space methods; Khepera robot; anticipative reinforcement learning; function approximation; generalisation; neural networks; state spaces; Australia; Books; Electric breakdown; Laboratories; Learning; Robots; Software algorithms; Software engineering; Space technology; State-space methods;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Information Processing, 2002. ICONIP '02. Proceedings of the 9th International Conference on
Print_ISBN :
981-04-7524-1
Type :
conf
DOI :
10.1109/ICONIP.2002.1202856
Filename :
1202856
Link To Document :
بازگشت