DocumentCode
72674
Title
Acceleration of Reinforcement Learning by Policy Evaluation Using Nonstationary Iterative Method
Author
Senda, K. ; Hattori, Saki ; Hishinuma, Toru ; Kohda, Tohru
Author_Institution
Dept. of Aeronaut. & Astronaut., Kyoto Univ., Kyoto, Japan
Volume
44
Issue
12
fYear
2014
fDate
Dec. 2014
Firstpage
2696
Lastpage
2705
Abstract
Typical methods for solving reinforcement learning problems iterate two steps, policy evaluation and policy improvement. This paper proposes algorithms for the policy evaluation to improve learning efficiency. The proposed algorithms are based on the Krylov Subspace Method (KSM), which is a nonstationary iterative method. The algorithms based on KSM are tens to hundreds times more efficient than existing algorithms based on the stationary iterative methods. Algorithms based on KSM are far more efficient than they have been generally expected. This paper clarifies what makes algorithms based on KSM makes more efficient with numerical examples and theoretical discussions.
Keywords
iterative methods; learning (artificial intelligence); KSM; Krylov subspace method; learning efficiency; nonstationary iterative method; policy evaluation; policy improvement; reinforcement learning problems; stationary iterative methods; Convergence; Eigenvalues and eigenfunctions; Equations; Iterative methods; Learning (artificial intelligence); Q-factor; Vectors; Nonstationary iterative method; policy evaluation; policy iteration; reinforcement learning;
fLanguage
English
Journal_Title
Cybernetics, IEEE Transactions on
Publisher
ieee
ISSN
2168-2267
Type
jour
DOI
10.1109/TCYB.2014.2313655
Filename
6786366
Link To Document