DocumentCode :
328281
Title :
Max-min propagation nets: learning by delta rule for the Chebyshev norm
Author :
Estevez, Pablo A. ; Okabe, Yoichi
Author_Institution :
Res. Center for Adv. Sci. & Technol., Tokyo Univ., Japan
Volume :
1
fYear :
1993
fDate :
25-29 Oct. 1993
Firstpage :
524
Abstract :
This paper introduces max-min propagation nets, which are composed of multiple-input maximum (max) and minimum (min) operators, besides linear weighted sum units. A learning algorithm based on gradient descent is derived for these networks. Two different criteria of error measurement are tested: the Chebyshev norm and the least squares norm. Weight-update rules for both criteria are deduced and implemented together with acceleration techniques to speed up convergence. A comparison of results on the parity problem is presented.
Keywords :
Chebyshev approximation; convergence of numerical methods; learning (artificial intelligence); minimax techniques; neural nets; Chebyshev norm; convergence; delta rule; error measurement; gradient descent; learning algorithm; least squares norm; linear weighted sum units; max-min propagation nets; maximum operators; minimum operators; parity problem; weight-update rules; Acceleration; Boolean functions; Chebyshev approximation; Convergence; Hardware; Least squares methods; Multi-layer neural network; Neural networks; Neurons; Testing;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 1993. IJCNN '93-Nagoya. Proceedings of 1993 International Joint Conference on
Print_ISBN :
0-7803-1421-2
Type :
conf
DOI :
10.1109/IJCNN.1993.713968
Filename :
713968
Link To Document :
بازگشت