DocumentCode
671664
Title
A new fast learning algorithm with promising global convergence capability for feed-forward neural networks
Author
Chi-Chung Cheung ; Sin-Chun Ng ; Lui, Andrew K. ; Xu, Sendren Sheng-Dong
Author_Institution
Dept. of Electron. & Inf. Eng., Hong Kong Polytech. Univ., Hong Kong, China
fYear
2013
fDate
4-9 Aug. 2013
Firstpage
1
Lastpage
6
Abstract
Backpropagation (BP) learning algorithm is the most widely used supervised learning technique that is extensively applied in the training of multi-layer feed-forward neural networks. Although many modifications of BP have been proposed to speed up the learning of the original BP, they seldom address the local minimum and the flat-spot problem. This paper proposes a new algorithm called Local-minimum and Flat-spot Problem Solver (LFPS) to solve these two problems. It uses a systematic approach to check whether a learning process is trapped by a local minimum or a flat-spot area, and then escape from it. Thus, a learning process using LFPS can keep finding an appropriate way to converge to the global minimum. The performance investigation shows that the proposed algorithm always converges in different learning problems (applications) whereas other popular fast learning algorithms sometimes give very poor global convergence capabilities.
Keywords
backpropagation; convergence; feedforward neural nets; BP learning algorithm; LFPS; backpropagation algorithm; global convergence capability; local-minimum and flat-spot problem solver; multilayer feedforward neural networks; systematic approach; Classification algorithms; Convergence; Databases; Educational institutions; Iris; Learning systems; Training;
fLanguage
English
Publisher
ieee
Conference_Titel
Neural Networks (IJCNN), The 2013 International Joint Conference on
Conference_Location
Dallas, TX
ISSN
2161-4393
Print_ISBN
978-1-4673-6128-6
Type
conf
DOI
10.1109/IJCNN.2013.6707006
Filename
6707006
Link To Document