DocumentCode :
1798444
Title :
Correntropy kernel temporal differences for reinforcement learning brain machine interfaces
Author :
Bae, Joonbum ; Sanchez Giraldo, Luis Gonzalo ; Principe, Jose C. ; Francis, Joseph T.
Author_Institution :
Dept. of Electr. & Comput. Eng., Univ. of Florida, Gainesville, FL, USA
fYear :
2014
fDate :
6-11 July 2014
Firstpage :
2713
Lastpage :
2717
Abstract :
This paper introduces a novel temporal difference algorithm to estimate a value function in reinforcement learning. This is a kernel adaptive system using a robust cost function called correntropy. We call this system correntropy kernel temporal differences (CKTD). This algorithm is integrated with Q-learning to find a proper policy (Q-learning via correntropy kernel temporal differences). The proposed method was tested with a synthetic problem, and its robustness under a changing policy was quantified. The same algorithm was applied to the decoding of a monkey´s neural states in a reinforcement learning brain machine interface (RLBMI) in a center-out reaching task. The results showed the potential advantage of the proposed algorithm in the RLBMI framework.
Keywords :
brain-computer interfaces; learning (artificial intelligence); CKTD; Q-learning; RLBMI framework; center-out reaching task; correntropy kernel temporal differences; kernel adaptive system; monkey neural states decoding; reinforcement learning brain machine interfaces; robust cost function; synthetic problem; value function estimation; Cost function; Decoding; Kernel; Learning (artificial intelligence); Monte Carlo methods; Robustness; Vectors;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks (IJCNN), 2014 International Joint Conference on
Conference_Location :
Beijing
Print_ISBN :
978-1-4799-6627-1
Type :
conf
DOI :
10.1109/IJCNN.2014.6889958
Filename :
6889958
Link To Document :
بازگشت