DocumentCode :
663819
Title :
Learning sequential tasks interactively from demonstrations and own experience
Author :
Grave, Kathrin ; Behnke, Sven
Author_Institution :
Dept. of Comput. Sci., Univ. of Bonn, Bonn, Germany
fYear :
2013
fDate :
3-7 Nov. 2013
Firstpage :
3237
Lastpage :
3243
Abstract :
Deploying robots to our day-to-day life requires them to have the ability to learn from their environment in order to acquire new task knowledge and to flexibly adapt existing skills to various situations. For typical real-world tasks, it is not sufficient to endow robots with a set of primitive actions. Rather, they need to learn how to sequence these in order to achieve a desired effect on their environment. In this paper, we propose an intuitive learning method for a robot to acquire sequences of motions by combining learning from human demonstrations and reinforcement learning. In every situation, our approach treats both ways of learning as alternative control flows to optimally exploit their strengths without inheriting their shortcomings. Using a Gaussian Process approximation of the state-action sequence value function, our approach generalizes values observed from demonstrated and autonomously generated action sequences to unknown inputs. This approximation is based on a kernel we designed to account for different representations of tasks and action sequences as well as inputs of variable length. From the expected deviation of value estimates, we devise a greedy exploration policy following a Bayesian optimization criterion that quickly converges learning to promising action sequences while protecting the robot from sequences with unpredictable outcome. We demonstrate the ability of our approach to efficiently learn appropriate action sequences in various situations on a manipulation task involving stacked boxes.
Keywords :
Bayes methods; Gaussian processes; approximation theory; learning (artificial intelligence); manipulators; motion control; Bayesian optimization criterion; Gaussian process approximation; action sequences; alternative control flow; greedy exploration policy; human demonstrations; intuitive learning method; manipulation task; real-world tasks; reinforcement learning; robots; sequential task learning; state-action sequence value function; Gaussian processes; Hidden Markov models; Kernel; Learning (artificial intelligence); Motion segmentation; Optimization;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Intelligent Robots and Systems (IROS), 2013 IEEE/RSJ International Conference on
Conference_Location :
Tokyo
ISSN :
2153-0858
Type :
conf
DOI :
10.1109/IROS.2013.6696816
Filename :
6696816
Link To Document :
بازگشت