DocumentCode
3520650
Title
How limited training data can allow a neural network to outperform an `optimal´ statistical classifier
Author
Niles, L. ; Silverman, Les Niles Harvey ; Tajchman, Gary ; Bush, Marcia
Author_Institution
Div. of Eng., Brown Univ., Providence, RI, USA
fYear
1989
fDate
23-26 May 1989
Firstpage
17
Abstract
Experiments comparing artificial neural network (ANN), k -nearest-neighbor (KNN), and Bayes´ rule with Gaussian distributions and maximum-likelihood estimation (BGM) classifiers were performed. Classifier error rate as a function of training set size was tested for synthetic data drawn from several different probability distributions. In cases where the true distributions were poorly modeled, ANN was significantly better than BGM. In some cases, ANN was also better than KNN. Similar experiments were performed on a voiced/unvoiced speech classification task. ANN had a lower error rate than KNN or BGM for all training set sizes, although BGM approached the ANN error rate as the training set became larger. It is concluded that there are pattern classification tasks in which an ANN is able to make better use of training data to achieve a lower error rate with a particular size training set
Keywords
neural nets; pattern recognition; speech recognition; Bayes´ rule; Gaussian distributions; artificial neural network; error rate; k-nearest-neighbor; limited training data; maximum-likelihood estimation; pattern classification tasks; probability distributions; statistical classifier; synthetic data; training set size; voiced/unvoiced speech classification; Artificial neural networks; Error analysis; Gaussian distribution; Maximum likelihood estimation; Neural networks; Pattern classification; Probability distribution; Speech; Testing; Training data;
fLanguage
English
Publisher
ieee
Conference_Titel
Acoustics, Speech, and Signal Processing, 1989. ICASSP-89., 1989 International Conference on
Conference_Location
Glasgow
ISSN
1520-6149
Type
conf
DOI
10.1109/ICASSP.1989.266352
Filename
266352
Link To Document