Title :
On variable sizes and sigmoid activation functions of multilayer perceptrons
Author :
Daqi, Gao ; Hua, Liu ; Changwu, Li
Author_Institution :
Dept. of Comput., East China Univ. of Sci. & Technol., Shanghai, China
Abstract :
This paper studies the influences of variable scales and sigmoid activation functions on the performances of multi-layer perceptrons. Generally speaking, it is not certainly suitable to normalize the input data or make the sizes of input variables in the range of [0.0, 1.0]. The viewpoint is explained in details according to the theory of support vector machine (SVM). The convergence and generalization abilities of multilayer perceptrons can be evidently improved by means of the following three methods: (A). Enlarge the sizes of the variable components in the range of [0.0, 3.0]. (B). Change the standard sigmoid activation function f(x)=(1+exp(-x))-1 into f(x)=3(1+exp(-x/3))-1. (C). Introduce the sum-of-squares weight term WTW into the error functions. The classification experiment shows that more than a learning round should be done and the perceptron with the best good generalization performance be held back.
Keywords :
convergence; generalisation (artificial intelligence); learning (artificial intelligence); multilayer perceptrons; pattern classification; support vector machines; transfer functions; SVM; convergence; error functions; input data normalization; multilayer perceptrons generalization abilities; sigmoid activation functions; sum-of-squares weight; support vector machine theory; variable components size; variable size activation functions; Bioreactors; Convergence; Error correction; Hardware; Laboratories; Large Hadron Collider; Multilayer perceptrons; Nonhomogeneous media; Paper technology; Support vector machines;
Conference_Titel :
Neural Networks, 2003. Proceedings of the International Joint Conference on
Print_ISBN :
0-7803-7898-9
DOI :
10.1109/IJCNN.2003.1223717