DocumentCode
2239742
Title
Convergence Analysis of Multiplicative Weight Noise Injection During Training
Author
Ho, Kevin ; Leung, Chi-Sing ; Sum, John ; Lau, Siu-chung
Author_Institution
Dept. of Comput. Sci. & Commun. Eng., Providence Univ., Sha-Lu, Taiwan
fYear
2010
fDate
18-20 Nov. 2010
Firstpage
358
Lastpage
365
Abstract
Injecting weight noise during training has been proposed for almost two decades as a simple technique to improve fault tolerance and generalization of a multilayer perceptron (MLP). However, little has been done regarding their convergence behaviors. Therefore, we presents in this paper the convergence proofs of two of these algorithms for MLPs. One is based on combining injecting multiplicative weight noise and weight decay (MWN-WD) during training. The other is based on combining injecting additive weight noise and weight decay (AWN-WD) during training. Let m be the number of hidden nodes of a MLP, a be the weight decay constant and Sb be the noise variance. It is showed that the convergence of MWN-WD algorithm is with probability one if a >; √(Sb)m. While the convergence of the AWN-WD algorithm is with probability one if a >; 0.
Keywords
fault tolerance; learning (artificial intelligence); multilayer perceptrons; probability; AWN-WD algorithm; MWN-WD algorithm; additive weight noise; convergence analysis; convergence behavior; convergence proof; fault tolerance; multilayer perceptron; multiplicative weight noise injection; noise variance; probability; training; weight decay; MLP; convergence; learning; weight noise;
fLanguage
English
Publisher
ieee
Conference_Titel
Technologies and Applications of Artificial Intelligence (TAAI), 2010 International Conference on
Conference_Location
Hsinchu City
Print_ISBN
978-1-4244-8668-7
Electronic_ISBN
978-0-7695-4253-9
Type
conf
DOI
10.1109/TAAI.2010.64
Filename
5695477
Link To Document