Title of article :
Parallel nonlinear optimization techniques for training neural networks
Author/Authors :
P.K.H.، Phua, نويسنده , , Ming، Daohua نويسنده ,
Issue Information :
روزنامه با شماره پیاپی سال 2003
Pages :
-145
From page :
146
To page :
0
Abstract :
In this paper, we propose the use of parallel quasi-Newton (QN) optimization techniques to improve the rate of convergence of the training process for neural networks. The parallel algorithms are developed by using the self-scaling quasi-Newton (SSQN) methods. At the beginning of each iteration, a set of parallel search directions is generated. Each of these directions is selectively chosen from a representative class of QN methods. Inexact line searches are then carried out to estimate the minimum point along each search direction. The proposed parallel algorithms are tested over a set of nine benchmark problems. Computational results show that the proposed algorithms outperform other existing methods, which are evaluated over the same set of test problems.
Journal title :
IEEE TRANSACTIONS ON NEURAL NETWORKS
Serial Year :
2003
Journal title :
IEEE TRANSACTIONS ON NEURAL NETWORKS
Record number :
62770
Link To Document :
بازگشت