DocumentCode
3428487
Title
A proposal of neural network architecture for non-linear function approximation
Author
Mizukami, Yoshiki ; Wakasa, Yuji ; Tanaka, Kanya
Author_Institution
Fac. of Eng., Yamaguchi Univ., Ube, Japan
Volume
4
fYear
2004
fDate
23-26 Aug. 2004
Firstpage
605
Abstract
In this paper, a neural network architecture for non-linear function approximation is proposed. We point out problems in non-linear function approximation with traditional neural networks, that is, difficulty in analyzing internal representation, no reproducibility in function approximation due to the random scheme for weight initialization, and the insufficient generalization ability in learning without enough samples. Based on these considerations, we suggest three main improvements. The first is the design of a sigmoidal function with localized derivative. The second is a deterministic scheme for weight initialization. The third is an updating rule for weight parameters. Simulation results show beneficial characteristics of our proposed method; low approximation error at the beginning of iterative calculation, smooth convergence of error and its improvement for difficulty in analyzing internal representation.
Keywords
approximation theory; neural net architecture; nonlinear functions; localized derivative; neural network architecture; nonlinear function approximation; sigmoidal function; updating rule; weight initialization; weight parameters; Analytical models; Approximation error; Convergence; Function approximation; Iterative methods; Linear approximation; Linearity; Neural networks; Proposals; Reproducibility of results;
fLanguage
English
Publisher
ieee
Conference_Titel
Pattern Recognition, 2004. ICPR 2004. Proceedings of the 17th International Conference on
ISSN
1051-4651
Print_ISBN
0-7695-2128-2
Type
conf
DOI
10.1109/ICPR.2004.1333845
Filename
1333845
Link To Document