DocumentCode :
892669
Title :
A generalized convergence theorem for neural networks
Author :
Bruck, Jehoshua ; Goodman, Joseph W.
Author_Institution :
Dept. of Electr. Eng., Stanford Univ., CA, USA
Volume :
34
Issue :
5
fYear :
1988
fDate :
9/1/1988 12:00:00 AM
Firstpage :
1089
Lastpage :
1092
Abstract :
A neural network model is presented in which each neuron performs a threshold logic function. The model always converges to a stable state when operating in a serial mode and to a cycle of length at most 2 when operating in a fully parallel mode. This property is the basis for the potential applications of the model, such as associative memory devices and combinatorial optimization. The two convergence theorems (for serial and fully parallel modes of operation) are reviewed, and a general convergence theorem is presented that unifies the two known cases. New relations between the neural network model and the problem of finding a minimum cut in a graph are obtained
Keywords :
combinatorial switching; convergence; neural nets; optimisation; associative memory devices; combinatorial optimisation; convergence theorems; fully parallel mode; neural networks; neuron; serial mode; stable state; threshold logic function; Associative memory; Computer networks; Convergence; Logic functions; Military computing; Multidimensional systems; Neural networks; Neurons; Performance evaluation; Symmetric matrices;
fLanguage :
English
Journal_Title :
Information Theory, IEEE Transactions on
Publisher :
ieee
ISSN :
0018-9448
Type :
jour
DOI :
10.1109/18.21239
Filename :
21239
Link To Document :
بازگشت