Title :
On using discretized Cohen-Grossberg node dynamics for model-free actor-critic neural learning in non-Markovian domains
Author :
Mizutani, Eiji ; Dreyfus, Stuart E.
Author_Institution :
Dept. of Comput. Sci., Nat. Tsing Hua Univ., Hsinchu, Taiwan
Abstract :
We describe how multi-stage non-Markovian decision problems can be solved using actor-critic reinforcement learning by assuming that a discrete version of Cohen-Grossberg node dynamics describes the node-activation computations of neural network (NN). Our NN is capable of rendering the process Markovian implicitly and automatically in a totally model-free fashion without learning by how much the state apace must be augmented so that the Markov property holds. This serves as an alternative to using Elman or Jordan-type function as a history memory in order to develop sensitivity to non-Markovian dependencies. We shall demonstrate our concept using a small-scale non-Markovian deterministic path problem, in which our actor-critic NN finds an optimal sequence of actions, although it needs much iteration due to the nature of neural model-free learning. This is, in spirit, a neuro-dynamic programming approach.
Keywords :
Markov processes; decision making; learning (artificial intelligence); neural nets; discretized Cohen-Grossberg node dynamics; model-free actor-critic neural learning; neural networks; nonMarkovian domains; optimal sequence of action; small-scale nonMarkovian deterministic path problem; Computer networks; Computer science; Dynamic programming; History; Learning; Neural networks; Neurons; Recurrent neural networks; Signal processing; State-space methods;
Conference_Titel :
Computational Intelligence in Robotics and Automation, 2003. Proceedings. 2003 IEEE International Symposium on
Print_ISBN :
0-7803-7866-0
DOI :
10.1109/CIRA.2003.1222053