• DocumentCode
    396728
  • Title

    Effect of regularization term upon fault tolerant training

  • Author

    Takase, Haruhiko ; Kita, Hidehiko ; Hayashi, Tetumine

  • Author_Institution
    Dept. of Electr. & Electron. Eng., Mie Univ., Japan
  • Volume
    2
  • fYear
    2003
  • fDate
    20-24 July 2003
  • Firstpage
    1048
  • Abstract
    To enhance fault tolerance of multi-layer neural networks, we proposed PAWMA (partially adaptive weight minimization approach). This method minimizes not only output error but also the sum of squares of weights (the regularization term). This method aims to decrease the number of connections whose faults strongly degrade the performance of MLNs (important connections). On the other hand, weight decay, which aims to eliminate unimportant connections, is base on the same idea. This method expects to keeping important connections and decaying unimportant connections. In this paper, we discuss about the contradiction between two effects of the regularization term. Through some experiment, we show that the difference between two effects is brought by the partially application of the regularization term.
  • Keywords
    backpropagation; fault tolerance; minimisation; multilayer perceptrons; fault tolerance; multilayer neural networks; partially adaptive weight minimization approach; regularization term; weight decay; Artificial neural networks; Degradation; Equations; Fault tolerance; Multi-layer neural network; Neural networks; Neurofeedback; Output feedback; Relays;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Neural Networks, 2003. Proceedings of the International Joint Conference on
  • ISSN
    1098-7576
  • Print_ISBN
    0-7803-7898-9
  • Type

    conf

  • DOI
    10.1109/IJCNN.2003.1223835
  • Filename
    1223835