Abstract :
The comparative performances of distributed and local neural networks for the speech recognition problem is investigated. Distributed networks´ hidden units use the signoid nonlinearity with global response. We have used the backpropagation rule with three error measures: mean square error, cross entropy, and combinational performance. The hidden units of local networks respond only to inputs in a certain local region in the input space. We used k-nearest neighbor (kNN), Gaussian-based kNN, learning vector quantization, and grow and learn methods. Phoneme recognition experiments were conducted using the /b,d,g,m,n,N/ set of the Japanese vocabulary for the speaker dependent case. Three criteria are considered for comparison: correct classification of the test set, network size, and learning time.
Keywords :
backpropagation; feedforward neural nets; speech recognition; vector quantisation; Gaussian-based k-nearest neighbor; Japanese phonemes; backpropagation rule; combinational performance; cross entropy; distributed neural classifiers; feedforward neural networks; hidden units; learning vector quantization; local neural classifiers; mean square error; signoid nonlinearity; speech recognition; Backpropagation; Entropy; Gaussian processes; Mean square error methods; Neural networks; Performance evaluation; Speech recognition; Testing; Vector quantization; Vocabulary;