Title :
Discrete recurrent neural networks for grammatical inference
Author :
Zeng, Zheng ; Goodman, Rodney M. ; Smyth, Padhraic
Author_Institution :
Dept. of Electr. Eng., California Inst. of Technol., Pasadena, CA, USA
fDate :
3/1/1994 12:00:00 AM
Abstract :
Describes a novel neural architecture for learning deterministic context-free grammars, or equivalently, deterministic pushdown automata. The unique feature of the proposed network is that it forms stable state representations during learning-previous work has shown that conventional analog recurrent networks can be inherently unstable in that they cannot retain their state memory for long input strings. The authors have previously introduced the discrete recurrent network architecture for learning finite-state automata. Here they extend this model to include a discrete external stack with discrete symbols. A composite error function is described to handle the different situations encountered in learning. The pseudo-gradient learning method (introduced in previous work) is in turn extended for the minimization of these error functions. Empirical trials validating the effectiveness of the pseudo-gradient learning method are presented, for networks both with and without an external stack. Experimental results show that the new networks are successful in learning some simple pushdown automata, though overfitting and non-convergent learning can also occur. Once learned, the internal representation of the network is provably stable; i.e., it classifies unseen strings of arbitrary length with 100% accuracy
Keywords :
context-free grammars; deterministic automata; learning (artificial intelligence); recurrent neural nets; composite error function; deterministic context-free grammars; deterministic pushdown automata; discrete external stack; discrete recurrent neural networks; grammatical inference; pseudo-gradient learning method; Laboratories; Law; Learning automata; Learning systems; Minimization methods; Neural networks; Propulsion; Recurrent neural networks; Space technology; Stability;
Journal_Title :
Neural Networks, IEEE Transactions on