Title :
Sensitivity-based inverse reinforcement learning
Author :
Zhaorong Tao ; Zhichao Chen ; Yanjie Li
Author_Institution :
Shenzhen Grad. Sch., Harbin Inst. of Technol., Shenzhen, China
Abstract :
Inverse reinforcement learning (IRL) is a process to obtain a potential reward function according to expert´s behavior. Then the optimal control policy is generated though some optimization theory, such as reinforcement learning, so that we can implement the imitation for expert´s behavior. In this paper, we consider the inverse reinforcement learning principle from t he point of performance sensitivity analysis. After that, we propose a novel inverse reinforcement learning analytical framework by analyzing the performance difference formula between expert´s policy and any other policies. This analytical framework extends the standard inverse reinforcement learning to the case that the reward function is related with both states and actions. At the same time, this framework provides a unified approach for IRL with the discount reward and the average reward in Markov decision process. Finally, the validity of corresponding results is verified under a grid world problem.
Keywords :
Markov processes; learning (artificial intelligence); optimal control; optimisation; sensitivity analysis; IRL; Markov decision process; expert behavior; expert policy; grid world problem; optimal control policy; optimization theory; performance difference formula; performance sensitivity analysis; reward function; sensitivity-based inverse reinforcement learning; Abstracts; Educational institutions; Electronic mail; Learning (artificial intelligence); Optimal control; Optimization; Sensitivity analysis; Inverse Reinforcement Learning; Performance sensitivity; Reward function;
Conference_Titel :
Control Conference (CCC), 2013 32nd Chinese
Conference_Location :
Xi´an