DocumentCode
2970967
Title
The learning convergence of CMAC in cyclic learning
Author
Yao, Shu ; Bo Zhang
Author_Institution
Dept. of Comput. Sci. & Technol., Tsinghua Univ., Beijing, China
Volume
3
fYear
1993
fDate
25-29 Oct. 1993
Firstpage
2583
Abstract
Discusses the learning convergence of the cerebellar model articulation controller (CMAC) in cyclic learning. The authors prove the following results. First, if the training samples are noiseless, the learning algorithm converges if and only if the learning rate is chosen from (0, 2). Second, when the training samples have noises, the learning algorithm will converge with probability one if the learning rate is dynamically decreased. Third, in the noise case, with a small but fixed learning rate ε the mean square error of the weight sequences generated by the CMAC learning algorithm will be bounded by O(ε). Some simulation experiments are carried out to test these results.
Keywords
cerebellar model arithmetic computers; convergence; learning (artificial intelligence); probability; CMAC; cerebellar model articulation controller; cyclic learning; learning convergence; mean square error; training samples; weight sequences; Associative memory; Backpropagation algorithms; Computer science; Convergence; Mean square error methods; Neural networks; Noise generators; Testing; Yttrium;
fLanguage
English
Publisher
ieee
Conference_Titel
Neural Networks, 1993. IJCNN '93-Nagoya. Proceedings of 1993 International Joint Conference on
Print_ISBN
0-7803-1421-2
Type
conf
DOI
10.1109/IJCNN.1993.714252
Filename
714252
Link To Document