Title :
Role of activation function on hidden units for sample recording in three-layer neural networks
Abstract :
It is shown that the k hidden units with asymptotic activation function are able to transfer any given k+1 different inputs to linearly independent GHUVs (generated hidden unit vectors) by properly setting weights and thresholds. The number of hidden units with the LIT (linearly independent transformation) capability for the polynomial activation function is limited by the order of polynomials. For analytic asymptotic activation functions and given different inputs, the LIT is a generic capability and a probability 1 capability in setting weights and thresholds randomly. It is a generic and a probability 1 property for any random input if the weight and threshold setting has LIT capability for some k+1 inputs. For three-layer nets with k hidden units, in which the activation function is asymptotic and the output layer is without activation function, they are sufficient to record k+1 arbitrary real samples. It is probability 0 to record k+2 random real samples if the activation is a unit step function. This is true for the sigmoid function in the case of associative memory. These conclusions lead to a scheme for understanding associative memory in the three-layer networks
Keywords :
content-addressable storage; learning systems; neural nets; LIT; analytic asymptotic activation functions; associative memory; asymptotic activation function; generated hidden unit vectors; hidden units; linearly independent GHUVs; linearly independent transformation; polynomial activation function; real samples; sigmoid function; three-layer nets; three-layer neural networks;
Conference_Titel :
Neural Networks, 1990., 1990 IJCNN International Joint Conference on
Conference_Location :
San Diego, CA, USA
DOI :
10.1109/IJCNN.1990.137548