Title :
On model selection and the disability of neural networks to decompose tasks
Author_Institution :
Inst. fur Neuroinformatik, Ruhr-Univ., Bochum, Germany
fDate :
6/24/1905 12:00:00 AM
Abstract :
A neural network with fixed topology can be regarded as a parametrization of functions, which decides on the correlations between functional variations when parameters are adapted. We propose an analysis, based on the differential geometry, that allows one to calculate these correlations. In practise, this describes how one response is unlearned while another is trained. Concerning conventional feed-forward neural networks we find that they generically introduce strong correlations, are predisposed to forgetting and inappropriate for task decomposition. Perspectives to solve these problems are discussed
Keywords :
differential geometry; feedforward neural nets; learning (artificial intelligence); topology; correlations; differential geometry; feedforward neural networks; forgetting; functional variations; learning; model parametrization; topology; Artificial neural networks; Data mining; Feature extraction; Feedforward neural networks; Feedforward systems; Geometry; Network topology; Neural networks;
Conference_Titel :
Neural Networks, 2002. IJCNN '02. Proceedings of the 2002 International Joint Conference on
Conference_Location :
Honolulu, HI
Print_ISBN :
0-7803-7278-6
DOI :
10.1109/IJCNN.2002.1005477