DocumentCode :
3251309
Title :
Why tanh: choosing a sigmoidal function
Author :
Kalman, Barry L. ; Kwasny, Stan C.
Author_Institution :
Dept. of Comput. Sci., Washington Univ., St. Louis, MO, USA
Volume :
4
fYear :
1992
fDate :
7-11 Jun 1992
Firstpage :
578
Abstract :
As hardware implementations of backpropagation and related training algorithms are anticipated, the choice of a sigmoidal function should be carefully justified. Attention should focus on choosing an activation function in a neural unit that exhibits the best properties for training. The author argues for the use of the hyperbolic tangent. While the exact shape of the sigmoidal makes little difference once the network is trained, it is shown that it possesses particular properties that make it appealing for use while training. By paying attention to scaling it is illustrated that tanh (1.5×) has the additional advantage of equalizing training over layers. This result can easily generalize to several standard sigmoidal functions commonly in use
Keywords :
backpropagation; neural nets; activation function; backpropagation; hyperbolic tangent; sigmoidal function; tanh; training algorithms; Computer science; Feedforward neural networks; Feedforward systems; Hardware; Intelligent systems; Kalman filters; Logistics; Neural networks; Shape; Telephony;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 1992. IJCNN., International Joint Conference on
Conference_Location :
Baltimore, MD
Print_ISBN :
0-7803-0559-0
Type :
conf
DOI :
10.1109/IJCNN.1992.227257
Filename :
227257
Link To Document :
بازگشت