DocumentCode
1242453
Title
Binary classification by stochastic neural nets
Author
Nádas, Arthur
Author_Institution
Res. Div., IBM Thomas J. Watson Res. Center, Yorktown Heights, NY, USA
Volume
6
Issue
2
fYear
1995
fDate
3/1/1995 12:00:00 AM
Firstpage
488
Lastpage
491
Abstract
We classify points in Rd (feature vector space) by functions related to feedforward artificial neural networks. These functions, dubbed “stochastic neural nets”, arise in a natural way from probabilistic as well as from statistical considerations. The probabilistic idea is to define a classifying bit locally by using the sign of a hidden state-dependent noisy linear function of the feature vector as a new (d+1)th coordinate of the vector. This (d+1)-dimensional distribution is approximated by a mixture distribution. The statistical idea is that the approximating mixtures, and hence the a posteriori class probability functions (stochastic neural nets) defined by them, can be conveniently trained either by maximum likelihood or by a Bayes criterion through the use of an appropriate expectation-maximization algorithm
Keywords
feedforward neural nets; pattern classification; probability; statistics; stochastic systems; Bayes criterion; approximating mixtures; binary classification; classifying bit; expectation-maximization algorithm; feature vector space; feedforward artificial neural networks; hidden state-dependent noisy linear function; local definition; maximum likelihood o; mixture distribution; point classification; probability functions; statistics; stochastic neural nets; Approximation error; Approximation methods; Artificial neural networks; Distribution functions; Gaussian noise; Neural networks; Probability; Stochastic processes; Stochastic resonance; Vectors;
fLanguage
English
Journal_Title
Neural Networks, IEEE Transactions on
Publisher
ieee
ISSN
1045-9227
Type
jour
DOI
10.1109/72.363484
Filename
363484
Link To Document