Title :
NeuralParse a neural model for parsing natural language
Author_Institution :
Rome Lab., NY, USA
Abstract :
Neural models for dealing with symbolic processing are in their infancy. Success thus far can be defined by the parsing of very simple phrases and a small set of words into small, fixed size frames. Many of these systems do not scale well as one increases the number of words or the phrase length. These models are limited with respect to the large number of epochs required to train and the error rates. We address the issue of training and present an analysis which provides a lower bound on the error rate. The approach investigates simple extensions to the basic learning algorithm and makes use of a closest neighbor algorithm for correctness
Keywords :
backpropagation; grammars; natural languages; neural nets; NeuralParse; closest neighbor algorithm; correctness; error rate; natural language; neural model; parsing; symbolic processing; training; Backpropagation; Concatenated codes; Data preprocessing; Encoding; Error analysis; Feeds; Instruments; Laboratories; Natural languages; Neural networks;
Conference_Titel :
Systems, Man, and Cybernetics, 1997. Computational Cybernetics and Simulation., 1997 IEEE International Conference on
Conference_Location :
Orlando, FL
Print_ISBN :
0-7803-4053-1
DOI :
10.1109/ICSMC.1997.635449