DocumentCode :
1807760
Title :
A learning method for synthesizing associative memory in neural networks
Author :
Kuroe, Yasuaki ; Koashi, Kenshu ; Hashimoto, Naoki ; Mori, Takehiro
Author_Institution :
Dept. of Electron. & Inf. Sci., Kyoto Inst. of Technol., Japan
Volume :
2
fYear :
1999
fDate :
36342
Firstpage :
798
Abstract :
The paper proposes a learning method for synthesizing associative memory in neural networks. The problem is formulated as determining the weights of the synaptic connections of neural networks such that, for any given set of desired memory vectors, each memory vector becomes an asymptotically stable equilibrium point of the network. We introduce a new architecture of neural networks, hybrid recurrent neural networks, in order to enhance the capability of implementing associative memories. An efficient learning method for synthesizing associative memories is proposed. The proposed method assures that all the memory vectors become asymptotically stable equilibrium points with the prescribed degree of stability. Synthesis examples are presented to demonstrate the applicability and performance of the proposed method
Keywords :
asymptotic stability; content-addressable storage; learning (artificial intelligence); neural net architecture; recurrent neural nets; associative memory; asymptotically stable equilibrium point; asymptotically stable equilibrium points; hybrid recurrent neural networks; learning method; memory vectors; synaptic connections; Artificial neural networks; Associative memory; Computer networks; Hopfield neural networks; Intelligent networks; Learning systems; Network synthesis; Neural networks; Neurons; Recurrent neural networks;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 1999. IJCNN '99. International Joint Conference on
Conference_Location :
Washington, DC
ISSN :
1098-7576
Print_ISBN :
0-7803-5529-6
Type :
conf
DOI :
10.1109/IJCNN.1999.831052
Filename :
831052
Link To Document :
بازگشت