• DocumentCode
    290811
  • Title

    Application of coding theory to neural net capacity

  • Author

    Al-Mashouq, Khalid A.

  • Author_Institution
    King Saud Univ., Riyadh, Saudi Arabia
  • fYear
    1994
  • fDate
    27 Jun-1 Jul 1994
  • Firstpage
    221
  • Abstract
    There are different definitions of neural net capacity. One of these definitions is what is called statistical pattern capacity. This capacity is defined as the average number of random patterns with random binary desired responses that can be “recognized” using a neural net. Another definition is what we may call worst case capacity. Consider a multilayer net with M input nodes and k output nodes. The “storage” capacity of this net is defined as the maximum number of input patterns for which the network can produce all possible output binary k-tuples. We use information theory, especially Shannon capacity theorem, to relate the neural net capacity to the channel capacity. As a practical example we demonstrate the effectiveness of error correcting codes to mitigate the imperfections of neural nets
  • Keywords
    channel capacity; error correction codes; feedforward neural nets; multilayer perceptrons; Shannon capacity theorem; channel capacity; coding theory; error correcting codes; information theory; input nodes; multilayer nets; neural net capacity; output nodes; random binary desired responses; random patterns; statistical pattern capacity; storage capacity; worst case capacity; Bandwidth; Communication channels; Error correction codes; Feedforward neural networks; Feedforward systems; Information theory; Multi-layer neural network; Neural networks; Nonhomogeneous media; Speech;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Information Theory, 1994. Proceedings., 1994 IEEE International Symposium on
  • Conference_Location
    Trondheim
  • Print_ISBN
    0-7803-2015-8
  • Type

    conf

  • DOI
    10.1109/ISIT.1994.394747
  • Filename
    394747