Title :
Overcoming the local-minimum problem in training multilayer perceptrons by gradual deconvexification
Author :
Lo, James Ting-Ho ; Yichuan Gui ; Yun Peng
Author_Institution :
Dept. of Math. & Stat., Univ. of Maryland, Baltimore, MD, USA
Abstract :
A method of training neural networks using the risk-averting error (RAE) criterion Jλ (w), which was presented in IJCNN 2001, has the capability to avoid nonglobal local minima, but suffers from a severe limitation on the magnitude of the risk-sensitivity index λ. To eliminating the limitation, an improved method using the normalized RAE (NRAE) Cλ (w) was proposed in ISNN 2012, but it requires a selection of a proper λ, whose range may be dependent on the application. A new training method called the gradual deconvexification (GDC) is proposed in this paper. It starts with a very large λ and gradually decreases it in the training process until a global minimum of Cλ (w) or a good generalization capability is achieved. GDC training method was tested on a large number of numerical examples and produced a very good result in each test.
Keywords :
mean square error methods; multilayer perceptrons; GDC training method; NRAE criterion; good generalization capability; gradual deconvexification; local-minimum problem; multilayer perceptrons; neural networks; normalized RAE criterion; risk-averting error criterion; Function approximation; Indexes; Registers; Training; Training data; Vectors;
Conference_Titel :
Neural Networks (IJCNN), The 2013 International Joint Conference on
Conference_Location :
Dallas, TX
Print_ISBN :
978-1-4673-6128-6
DOI :
10.1109/IJCNN.2013.6706796