DocumentCode :
288594
Title :
Backpropagation without weight transport
Author :
Kolen, John F. ; Pollack, Jordan B.
Author_Institution :
Lab. for Artificial Intelligence Res., Ohio State Univ., Columbus, OH, USA
Volume :
3
fYear :
1994
fDate :
27 Jun-2 Jul 1994
Firstpage :
1375
Abstract :
In backpropagation, connection weights are used to both compute node activations and error gradient for hidden units. Grossberg (1987) has argued that the dual use of the same synaptic connections (weight transport) constitutes a bidirectional flow of information through synapses, which is biologically implausable. In this paper we formally and empirically demonstrate the feasibility of an architecture equivalent to backpropagation, but without the assumption of weight transport. Through coordinated training with weight decay, a reciprocal layer of weights evolves into a copy of the forward connections and acts as the conduit for backward flowing corrective information. Examination of the networks trained with dual weights suggests that functional synchronization, and not weight synchronization, is crucial to the operation of backpropagation methods
Keywords :
backpropagation; neural net architecture; neural nets; synchronisation; backpropagation; backward flowing corrective information; connection weights; coordinated training; error gradient; functional synchronization; hidden units; node activations; synaptic connections; weight decay; Backpropagation;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 1994. IEEE World Congress on Computational Intelligence., 1994 IEEE International Conference on
Conference_Location :
Orlando, FL
Print_ISBN :
0-7803-1901-X
Type :
conf
DOI :
10.1109/ICNN.1994.374486
Filename :
374486
Link To Document :
بازگشت