DocumentCode :
276650
Title :
Linear neural networks which minimize the output variance
Author :
Palmieri, Francesco ; Zhu, Jie
Author_Institution :
Dept. of Electr. & Syst. Eng., Connecticut Univ., Storrs, CT, USA
Volume :
i
fYear :
1991
fDate :
8-14 Jul 1991
Firstpage :
791
Abstract :
The authors analyze constrained linear architectures which learn according to Hebb´s rule to minimize the output energy. They study the conditions under which such networks act as decorrelating (square-root) filters. In particular, it is shown how constrained architectures can decorrelate efficiently by simply using Hebb´s rule. The authors extend the analysis to networks with arbitrary interconnections. The purpose is the design of useful architectures and the understanding of the functionality of patterns of connectivity observed in biological systems. The authors deal only with linear neurons performing simple linear combinations. The authors restrict attention to decorrelating networks which do not use the output variance to compress the input space into a new space with a smaller number of dimensions
Keywords :
filtering and prediction theory; learning systems; neural nets; Hebb´s rule; biological systems; connectivity; constrained linear architectures; decorrelating filters; decorrelating networks; linear neural nets; linear neurons; output variance minimisation; square root filters; Analysis of variance; Convergence; Decorrelation; Least squares methods; Neural networks; Neurons; Nonlinear filters; Power engineering and energy; Systems engineering and theory; Vectors;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 1991., IJCNN-91-Seattle International Joint Conference on
Conference_Location :
Seattle, WA
Print_ISBN :
0-7803-0164-1
Type :
conf
DOI :
10.1109/IJCNN.1991.155279
Filename :
155279
Link To Document :
بازگشت