DocumentCode :
3590751
Title :
Learning regular languages via recurrent higher-order neural networks
Author :
Tanaka, Ken ; Kumazawa, Itsuo
Author_Institution :
Fac. of Eng., Niigata Univ., Japan
Volume :
2
fYear :
1996
Firstpage :
1378
Abstract :
Learning regular languages is accomplished by the acquisition of finite state automata. In order for a neural network to acquire an arbitrary FSA, the network must first have a representation for every FSA state, and furthermore be able to realize an arbitrary FSA state transfer function. We show that if the network model of Giles et al. (1992) represents each FSA state using local state representation, then it can realize any FSA state transfer function. However, this may be difficult to acquire by learning, and is a reason why the model is not necessarily successful at learning some regular languages. In order to overcome this problem we propose the recurrent higher-order neural network (RHON). We show the order of connections sufficient to realize any FSA state transfer function regardless of the networks representation of states. After deriving the learning algorithm, we show the learning superiority of RHON to the model of Giles et al. through computer simulation
Keywords :
finite automata; formal languages; learning (artificial intelligence); recurrent neural nets; finite state automata; local state representation; recurrent higher-order neural networks; regular languages; state transfer function; Gradient methods; Learning automata; NP-hard problem; Neural networks; Recurrent neural networks; Transfer functions;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 1996., IEEE International Conference on
Print_ISBN :
0-7803-3210-5
Type :
conf
DOI :
10.1109/ICNN.1996.549100
Filename :
549100
Link To Document :
بازگشت