DocumentCode :
303305
Title :
A neural model of sequential memory
Author :
Wang, DeLiang ; Yuwono, Budi
Author_Institution :
Dept. of Comput. & Inf. Sci., Ohio State Univ., Columbus, OH, USA
Volume :
2
fYear :
1996
fDate :
3-6 Jun 1996
Firstpage :
834
Abstract :
A neural model for temporal pattern generation is analyzed for learning multiple complex sequences in a sequential manner. The network exhibits a degree of interference when new sequences are acquired. It is proven that the model is capable of incrementally learning a finite number of complex sequences. The model is then evaluated with a large set of highly correlated sequences. While the number of intact sequences increases linearly, the amount of retraining due to interference appears to be independent of the size of existing memory. The idea of chunking helps to substantially reduce the amount of retraining in sequential learning. The network investigated here constitutes an effective sequential memory
Keywords :
learning (artificial intelligence); neural nets; chunking; multiple complex sequences; neural model; retraining; sequential memory; temporal pattern generation; Artificial intelligence; Associative memory; Biological neural networks; Cognitive science; Context modeling; Information analysis; Information science; Interference; Multilayer perceptrons; Pattern analysis;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 1996., IEEE International Conference on
Conference_Location :
Washington, DC
Print_ISBN :
0-7803-3210-5
Type :
conf
DOI :
10.1109/ICNN.1996.549005
Filename :
549005
Link To Document :
بازگشت