• DocumentCode
    3692653
  • Title

    The optimistic exploration value function

  • Author

    Michal Gregor;Juraj Spalek

  • Author_Institution
    Department of Control and Information Systems, University of Ž
  • fYear
    2015
  • Firstpage
    119
  • Lastpage
    123
  • Abstract
    The paper presents an approach that uses optimistic initialization and scalarized multi-objective learning to facilitate exploration in the context of model-free reinforcement learning. In contrast to existing optimistic intialization approaches, the approach introduces an extra value function, which is initialized optimistically and then updated using a zero reward function. Linear or Chebyshev scalarization is then used to compound this function with the standard task-related value function, thus forming an exploration policy. The paper concludes with evaluation of the approach on a benchmark task.
  • Keywords
    Chebyshev approximation
  • Publisher
    ieee
  • Conference_Titel
    Intelligent Engineering Systems (INES), 2015 IEEE 19th International Conference on
  • Type

    conf

  • DOI
    10.1109/INES.2015.7329650
  • Filename
    7329650