Title : 
Differential learning leads to efficient neural network classifiers
         
        
            Author : 
Hampshire, J.B., II ; Kumar, B. K K Vjaya
         
        
            Author_Institution : 
Dept. of Electr. & Comput. Eng., Carnegie Mellon Univ., Pittsburgh, PA, USA
         
        
        
        
        
        
            Abstract : 
The authors outline a differential theory of learning for statistical pattern classification. The theory is based on classification figure-of-merit (CFM) objective functions, described by J. P. Hampshire II and A. H. Waibel (IEEE Trans. Neural Netw. vol.1, no.2, p.216-218, June 1990). They give the proof that differential learning is efficient, requiring the least classifier complexity and the smallest training sample size necessary to achieve Bayesian (i.e., minimum error) discrimination. A practical application of the theory is included in which a simple differentially trained linear neural network classifier discriminations handwritten digits of the AT&T DB1 database with a 1.3% error rate. This error rate is less than one half of the best previous result for a linear classifier on this optical character recognition (OCR) task.<>
         
        
            Keywords : 
Bayes methods; computational complexity; learning (artificial intelligence); neural nets; optical character recognition; Bayesian discrimination; IEEE Trans. Neural Netw.; classifier complexity; differential theory of learning; error rate; handwritten digits; linear neural network classifier; objective functions; optical character recognition; statistical pattern classification; training sample size;
         
        
        
        
            Conference_Titel : 
Acoustics, Speech, and Signal Processing, 1993. ICASSP-93., 1993 IEEE International Conference on
         
        
            Conference_Location : 
Minneapolis, MN, USA
         
        
        
            Print_ISBN : 
0-7803-7402-9
         
        
        
            DOI : 
10.1109/ICASSP.1993.319193