DocumentCode :
3309533
Title :
Reducing computations in incremental learning for feedforward neural network with long-term memory
Author :
Kobyashi, M. ; Zamani, Anuar ; Ozawa, Seiichi ; Abe, Shigeo
Author_Institution :
Graduate Sch. of Sci. & Technol., Kobe Univ., Japan
Volume :
3
fYear :
2001
fDate :
2001
Firstpage :
1989
Abstract :
When neural networks are trained incrementally, input-output relationships that are trained formerly tend to be collapsed by the learning of new training data. This phenomenon is called “interference”. To suppress the interference, we have proposed an incremental learning system (called RAN-LTM), in which long-term memory (LTM) is introduced into a resource allocating network (RAN). Since RAN-LTM needs to train not only new data but also some LTM data to suppress the interference, if many LTM data are retrieved large computations are required. Therefore, it is important to design appropriate procedures for producing and retrieving LTM data in RAN-LTM. In the paper, these procedures in the previous version of RAN-LTM are improved. In simulations, the improved RAN-LTM is applied to the approximation of a one-dimensional function, and the approximation error and the training speed are evaluated as compared with RAN and the previous RAN-LTM
Keywords :
feedforward neural nets; learning (artificial intelligence); RAN-LTM; approximation error; computations reduction; feedforward neural network; incremental learning; input-output relationships; long-term memory; resource allocating network; training speed; Buffer storage; Computer networks; Feedforward neural networks; Information retrieval; Intelligent networks; Interference suppression; Learning systems; Neural networks; Radio access networks; Training data;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 2001. Proceedings. IJCNN '01. International Joint Conference on
Conference_Location :
Washington, DC
ISSN :
1098-7576
Print_ISBN :
0-7803-7044-9
Type :
conf
DOI :
10.1109/IJCNN.2001.938469
Filename :
938469
Link To Document :
بازگشت