Title :
Plan-based reward shaping for reinforcement learning
Author :
Grzes, Marek ; Kudenko, Daniel
Author_Institution :
Dept. of Comput. Sci., Univ. of York, York
Abstract :
Reinforcement learning, while being a highly popular learning technique for agents and multi-agent systems, has so far encountered difficulties when applying it to more complex domains due to scaling-up problems. This paper focuses on the use of domain knowledge to improve the convergence speed and optimality of various RL techniques. Specifically, we propose the use of high-level STRIPS operator knowledge in reward shaping to focus the search for the optimal policy. Empirical results show that the plan-based reward shaping approach outperforms other RL techniques, including alternative manual and MDP-based reward shaping when it is used in its basic form. We show that MDP-based reward shaping may fail and successful experiments with STRIPS-based shaping suggest modifications which can overcome encountered problems. The STRIPS-based method we propose allows expressing the same domain knowledge in a different way and the domain expert can choose whether to define an MDP or STRIPS planning task. We also evaluate the robustness of the proposed STRIPS-based technique to errors in the plan knowledge.
Keywords :
Markov processes; convergence; learning (artificial intelligence); multi-agent systems; planning (artificial intelligence); Markov decision process planning; convergence speed; high-level STRIPS operator knowledge planning; multiagent system; optimal policy; plan-based reward shaping approach; reinforcement learning; Convergence; Intelligent agent; Intelligent systems; Learning; Manuals; Multiagent systems; Robustness; Shape control; State feedback; Strips; Reinforcement learning; STRIPS; reward shaping; symbolic planning;
Conference_Titel :
Intelligent Systems, 2008. IS '08. 4th International IEEE Conference
Conference_Location :
Varna
Print_ISBN :
978-1-4244-1739-1
Electronic_ISBN :
978-1-4244-1740-7
DOI :
10.1109/IS.2008.4670492