DocumentCode :
303202
Title :
Sparse initial topologies for high order perceptrons
Author :
Pol, A. De ; Thimm, G. ; Fiesler, E.
Author_Institution :
IDIAP, Martigny, Switzerland
Volume :
1
fYear :
1996
fDate :
3-6 Jun 1996
Firstpage :
84
Abstract :
High order neural networks are more powerful than first order neural networks. Their main drawback is the large number of connections that they entail in their default form, which is fully interlayer connected. The methods presented here produce initial topologies that only use a very small fraction of the possible connections, yet provide a good framework for successful training of high order perceptrons. There initial topologies can be refined using ontogenic techniques that modify the topology during the training of the network. The methods are based on approximating real valued data by Boolean data, which is used as a basis for constructing the network structure. The methods are evaluated for their effectiveness in reducing the network size by comparing them to fully interlayer connected high order perceptrons and their performance is evaluated by testing the generalization capabilities of the resulting networks
Keywords :
Boolean functions; approximation theory; generalisation (artificial intelligence); network topology; perceptrons; Boolean logic; connectionism; data approximation; generalization; high order perceptrons; ontogenic neural networks; sparse initial topology; Boolean functions; Combinatorial mathematics; Equations; Explosions; Network topology; Neural networks; Neurons; Polynomials; Splicing; Testing;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 1996., IEEE International Conference on
Conference_Location :
Washington, DC
Print_ISBN :
0-7803-3210-5
Type :
conf
DOI :
10.1109/ICNN.1996.548871
Filename :
548871
Link To Document :
بازگشت