DocumentCode
3312673
Title
Using dimensionality reduction to exploit constraints in reinforcement learning
Author
Bitzer, Sebastian ; Howard, Matthew ; Vijayakumar, Sethu
Author_Institution
Sch. of Inf., Univ. of Edinburgh, Edinburgh, UK
fYear
2010
fDate
18-22 Oct. 2010
Firstpage
3219
Lastpage
3225
Abstract
Reinforcement learning in the high-dimensional, continuous spaces typical in robotics, remains a challenging problem. To overcome this challenge, a popular approach has been to use demonstrations to find an appropriate initialisation of the policy in an attempt to reduce the number of iterations needed to find a solution. Here, we present an alternative way to incorporate prior knowledge from demonstrations of individual postures into learning, by extracting the inherent problem structure to find an efficient state representation. In particular, we use probabilistic, nonlinear dimensionality reduction to capture latent constraints present in the data. By learning policies in the learnt latent space, we are able to solve the planning problem in a reduced space that automatically satisfies task constraints. As shown in our experiments, this reduces the exploration needed and greatly accelerates the learning. We demonstrate our approach for learning a bimanual reaching task on the 19-DOF KHR-1HV humanoid.
Keywords
humanoid robots; learning (artificial intelligence); path planning; robot kinematics; 19-DOF KHR-1HV humanoid; continuous spaces; dimensionality reduction; latent constraints; learning policies; nonlinear dimensionality reduction; planning problem; probabilistic reduction; reinforcement learning; robotics; state representation;
fLanguage
English
Publisher
ieee
Conference_Titel
Intelligent Robots and Systems (IROS), 2010 IEEE/RSJ International Conference on
Conference_Location
Taipei
ISSN
2153-0858
Print_ISBN
978-1-4244-6674-0
Type
conf
DOI
10.1109/IROS.2010.5650243
Filename
5650243
Link To Document