Title :
A general probabilistic formulation for neural classifiers
Author_Institution :
Dept. of Comput. Sci., Maryland Univ., Baltimore, MD
Abstract :
We use partial likelihood (PL) theory to introduce a general probabilistic framework for the design and analysis of neural classifiers. The formulation allows for the training samples used in the design to have correlations in time, and for use of a wide range of neural network probability models including recurrent structures. We use PL theory to establish a fundamental information-theoretic connection, show the equivalence of likelihood maximization and relative entropy minimization, without making the common assumptions of independent training samples and true distribution information. Large sample optimality properties of PL can also be established under mild regularity conditions which allows adaptive-structure and robust classifier designs by using modified likelihood functions and information-theoretic criteria
Keywords :
feedforward neural nets; learning (artificial intelligence); minimum entropy methods; pattern classification; probability; recurrent neural nets; adaptive-structure classifier; general probabilistic formulation; information-theoretic connection; information-theoretic criteria; large sample optimality properties; likelihood maximization; mild regularity conditions; modified likelihood functions; neural classifiers; neural network probability models; partial likelihood theory; recurrent structures; relative entropy minimization; robust classifier; training samples; Computer science; Entropy; Maximum likelihood estimation; Mean square error methods; Neural networks; Parameter estimation; Probability; Recurrent neural networks; Robustness; Supervised learning;
Conference_Titel :
Neural Networks for Signal Processing VIII, 1998. Proceedings of the 1998 IEEE Signal Processing Society Workshop
Conference_Location :
Cambridge
Print_ISBN :
0-7803-5060-X
DOI :
10.1109/NNSP.1998.710644