DocumentCode
186269
Title
Inverse reinforcement learning using Dynamic Policy Programming
Author
Uchibe, Eiji ; Doya, Kenji
Author_Institution
Neural Comput. Unit, Okinawa Inst. of Sci. & Technol. Grad. Univ., Okinawa, Japan
fYear
2014
fDate
13-16 Oct. 2014
Firstpage
222
Lastpage
228
Abstract
This paper proposes a novel model-free inverse reinforcement learning method based on density ratio estimation under the framework of Dynamic Policy Programming. We show that the logarithm of the ratio between the optimal policy and the baseline policy is represented by the state-dependent cost and the value function. Our proposal is to use density ratio estimation methods to estimate the density ratio of policies and the least squares method with regularization to estimate the state-dependent cost and the value function that satisfies the relation. Our method can avoid computing the integral such as evaluating the partition function. A simple numerical simulation of a grid world navigation, a car driving, and a pendulum swing-up shows its superiority over conventional methods.
Keywords
dynamic programming; estimation theory; learning (artificial intelligence); least squares approximations; car driving; density ratio estimation method; dynamic policy programming; grid world navigation; inverse reinforcement learning; least squares method; pendulum swing-up; state-dependent cost; value function; Cost function; Estimation; Learning (artificial intelligence); Mathematical model; Navigation; Trajectory; Vectors;
fLanguage
English
Publisher
ieee
Conference_Titel
Development and Learning and Epigenetic Robotics (ICDL-Epirob), 2014 Joint IEEE International Conferences on
Conference_Location
Genoa
Type
conf
DOI
10.1109/DEVLRN.2014.6982985
Filename
6982985
Link To Document