Title :
Why tanh: choosing a sigmoidal function
Author :
Kalman, Barry L. ; Kwasny, Stan C.
Author_Institution :
Dept. of Comput. Sci., Washington Univ., St. Louis, MO, USA
Abstract :
As hardware implementations of backpropagation and related training algorithms are anticipated, the choice of a sigmoidal function should be carefully justified. Attention should focus on choosing an activation function in a neural unit that exhibits the best properties for training. The author argues for the use of the hyperbolic tangent. While the exact shape of the sigmoidal makes little difference once the network is trained, it is shown that it possesses particular properties that make it appealing for use while training. By paying attention to scaling it is illustrated that tanh (1.5×) has the additional advantage of equalizing training over layers. This result can easily generalize to several standard sigmoidal functions commonly in use
Keywords :
backpropagation; neural nets; activation function; backpropagation; hyperbolic tangent; sigmoidal function; tanh; training algorithms; Computer science; Feedforward neural networks; Feedforward systems; Hardware; Intelligent systems; Kalman filters; Logistics; Neural networks; Shape; Telephony;
Conference_Titel :
Neural Networks, 1992. IJCNN., International Joint Conference on
Conference_Location :
Baltimore, MD
Print_ISBN :
0-7803-0559-0
DOI :
10.1109/IJCNN.1992.227257