Title :
On the number of memories that can be perfectly stored in a neural net with Hebb weights
Author_Institution :
Dept. of Math., Rutgers Univ., New Brunswick, NJ, USA
fDate :
1/1/1989 12:00:00 AM
Abstract :
Let {wij} be the weights of the connections of a neural network with n nodes, calculated from m data vectors v1, ···, vm in {1,-1}n, according to the Hebb rule. The author proves that if m is not too large relative to n and the vk are random, then the wij constitute, with high probability, a perfect representation of the vk in the sense that the v k are completely determined by the wij up to their sign. The conditions under which this is established turn out to be less restrictive than those under which it has been shown that the vk can actually be recovered by letting the network evolve until equilibrium is attained. In the specific case where the entries of the vk are independent and equal to 1 or -1 with probability 1/2, the condition on m is that m should not exceed n/0.7 log n
Keywords :
content-addressable storage; information theory; neural nets; Hebb weights; associative memory; neural net; Associative memory; Intelligent networks; Mathematics; Neural networks; Neurons;
Journal_Title :
Information Theory, IEEE Transactions on