DocumentCode :
423536
Title :
Context discerning multifunction networks: reformulating fixed weight neural networks
Author :
Santiago, Roberto A.
Author_Institution :
NW Comput. Intelligence Lab., Portland State Univ., OR, USA
Volume :
1
fYear :
2004
fDate :
25-29 July 2004
Lastpage :
194
Abstract :
Research in recurrent neural networks has produced a genre of networks referred to as fixed weight neural networks (FWNNs) which have the ability to adapt without changing explicit weights. FWNNs are unique in that they adapt their processing based on the spatiotemporal characteristics of the incoming signal without need for weight change. As a result, a single FWNN is able to model and control many families of disparate systems without weight changes. FWNNs pose an interesting model for contextual memory in neural systems. The work reported takes a FWNN, decomposes it and analyzes its internal workings. Using new insight, FWNNs are reformulated into a simpler structure, context discerning multifunction networks (CDMN).
Keywords :
learning (artificial intelligence); recurrent neural nets; context discerning multifunction network; contextual memory; fixed weight neural network reformulation; neural systems; recurrent neural networks; spatiotemporal characteristics; Computational intelligence; Context modeling; Electronic mail; Feedforward neural networks; Machine learning; Neural networks; Recurrent neural networks; Signal processing; Spatiotemporal phenomena;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 2004. Proceedings. 2004 IEEE International Joint Conference on
ISSN :
1098-7576
Print_ISBN :
0-7803-8359-1
Type :
conf
DOI :
10.1109/IJCNN.2004.1379896
Filename :
1379896
Link To Document :
بازگشت