DocumentCode :
3517365
Title :
Learning objective functions for manipulation
Author :
Kalakrishnan, Mrinal ; Pastor, Peter ; Righetti, Ludovic ; Schaal, Stefan
Author_Institution :
CLMC Lab., Univ. of Southern California, Los Angeles, CA, USA
fYear :
2013
fDate :
6-10 May 2013
Firstpage :
1331
Lastpage :
1336
Abstract :
We present an approach to learning objective functions for robotic manipulation based on inverse reinforcement learning. Our path integral inverse reinforcement learning algorithm can deal with high-dimensional continuous state-action spaces, and only requires local optimality of demonstrated trajectories. We use L1 regularization in order to achieve feature selection, and propose an efficient algorithm to minimize the resulting convex objective function. We demonstrate our approach by applying it to two core problems in robotic manipulation. First, we learn a cost function for redundancy resolution in inverse kinematics. Second, we use our method to learn a cost function over trajectories, which is then used in optimization-based motion planning for grasping and manipulation tasks. Experimental results show that our method outperforms previous algorithms in high-dimensional settings.
Keywords :
learning (artificial intelligence); manipulator kinematics; path planning; convex objective function; feature selection; high-dimensional continuous state-action spaces; inverse kinematics; learning objective functions; optimization-based motion planning; path integral inverse reinforcement learning algorithm; robotic manipulation; Cost function; Joints; Kinematics; Learning (artificial intelligence); Robots; Trajectory;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Robotics and Automation (ICRA), 2013 IEEE International Conference on
Conference_Location :
Karlsruhe
ISSN :
1050-4729
Print_ISBN :
978-1-4673-5641-1
Type :
conf
DOI :
10.1109/ICRA.2013.6630743
Filename :
6630743
Link To Document :
بازگشت