Title :
Inverse reinforcement learning using Dynamic Policy Programming
Author :
Uchibe, Eiji ; Doya, Kenji
Author_Institution :
Neural Comput. Unit, Okinawa Inst. of Sci. & Technol. Grad. Univ., Okinawa, Japan
Abstract :
This paper proposes a novel model-free inverse reinforcement learning method based on density ratio estimation under the framework of Dynamic Policy Programming. We show that the logarithm of the ratio between the optimal policy and the baseline policy is represented by the state-dependent cost and the value function. Our proposal is to use density ratio estimation methods to estimate the density ratio of policies and the least squares method with regularization to estimate the state-dependent cost and the value function that satisfies the relation. Our method can avoid computing the integral such as evaluating the partition function. A simple numerical simulation of a grid world navigation, a car driving, and a pendulum swing-up shows its superiority over conventional methods.
Keywords :
dynamic programming; estimation theory; learning (artificial intelligence); least squares approximations; car driving; density ratio estimation method; dynamic policy programming; grid world navigation; inverse reinforcement learning; least squares method; pendulum swing-up; state-dependent cost; value function; Cost function; Estimation; Learning (artificial intelligence); Mathematical model; Navigation; Trajectory; Vectors;
Conference_Titel :
Development and Learning and Epigenetic Robotics (ICDL-Epirob), 2014 Joint IEEE International Conferences on
Conference_Location :
Genoa
DOI :
10.1109/DEVLRN.2014.6982985