• DocumentCode
    2949997
  • Title

    S-learning: a reinforcement learning method without parameter tuning

  • Author

    Chen, Hown-Wen ; Soo, Von-Wun

  • Author_Institution
    Dept. of Comput. Sci., Nat. Tsing Hua Univ., Hsinchu, Taiwan
  • Volume
    1
  • fYear
    1993
  • fDate
    25-29 Oct. 1993
  • Firstpage
    557
  • Abstract
    Discussed issues raised in two reinforcement learning algorithms, adaptive heuristic critic (AHC) and Q-learning. Aspects considered are convergence, parameter tuning, over-training, and computational and storage efficiency. Two new reinforcement learning mechanisms are proposed: S-learning (step-learning) and S&AHC learning. Particularly, the representation of the final cost map formed by S-learning series can be explicitly interpreted as the number of minimum movements to the goal state from each individual state. In addition, an adaptive (incremental) S-learning was proposed which incorporated S-learning and the technique of incremental learning to facilitate the practical implementation of neural reinforcement learning. All of S-learning series showed promising performances in exploring Sutton´s task (1991) of navigating in a maze.
  • Keywords
    computational complexity; heuristic programming; learning (artificial intelligence); AHC; Q-learning; S&AHC learning; S-learning; adaptive heuristic critic; computational efficiency; convergence; neural reinforcement learning; over-training; parameter tuning; reinforcement learning method; step-learning; storage efficiency; Biological control systems; Biological system modeling; Computer science; Convergence; Costs; Delay; Game theory; Learning; Navigation; State estimation;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Neural Networks, 1993. IJCNN '93-Nagoya. Proceedings of 1993 International Joint Conference on
  • Print_ISBN
    0-7803-1421-2
  • Type

    conf

  • DOI
    10.1109/IJCNN.1993.713976
  • Filename
    713976