DocumentCode :
3692653
Title :
The optimistic exploration value function
Author :
Michal Gregor;Juraj Spalek
Author_Institution :
Department of Control and Information Systems, University of Ž
fYear :
2015
Firstpage :
119
Lastpage :
123
Abstract :
The paper presents an approach that uses optimistic initialization and scalarized multi-objective learning to facilitate exploration in the context of model-free reinforcement learning. In contrast to existing optimistic intialization approaches, the approach introduces an extra value function, which is initialized optimistically and then updated using a zero reward function. Linear or Chebyshev scalarization is then used to compound this function with the standard task-related value function, thus forming an exploration policy. The paper concludes with evaluation of the approach on a benchmark task.
Keywords :
Chebyshev approximation
Publisher :
ieee
Conference_Titel :
Intelligent Engineering Systems (INES), 2015 IEEE 19th International Conference on
Type :
conf
DOI :
10.1109/INES.2015.7329650
Filename :
7329650
Link To Document :
بازگشت